<?xml version="1.0" encoding="UTF-8" standalone="no"?><!--Generated by Site-Server v@build.version@ (http://www.squarespace.com) on Thu, 16 Apr 2026 08:37:14 GMT
--><rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:media="http://www.rssboard.org/media-rss" xmlns:wfw="http://wellformedweb.org/CommentAPI/" version="2.0"><channel><title>Jason Burke's Blog</title><link>https://www.jasonburke.online/blog/</link><lastBuildDate>Thu, 26 Mar 2026 12:44:27 +0000</lastBuildDate><language>en-US</language><generator>Site-Server v@build.version@ (http://www.squarespace.com)</generator><description>Perspectives on AI, Analytics, and Transformation</description><item><title>AI Is Not a Technology Problem — It’s an Operating Model Transformation</title><dc:creator>Jason Burke</dc:creator><pubDate>Tue, 24 Mar 2026 19:08:00 +0000</pubDate><link>https://www.jasonburke.online/blog/ai-is-not-a-technology-problem-its-an-operating-model-transformation</link><guid isPermaLink="false">674a65c6ba0c721888a63309:674a6885532ea52b2bdb097c:69c2dfcf13b45f618e7297ef</guid><description><![CDATA[Most organizations treat AI as a technology initiative. In healthcare and 
life sciences, real value comes from transforming how institutions operate.]]></description><content:encoded><![CDATA[<p data-rte-preserve-empty="true">Boards are asking about it. Executives are funding it. Teams are experimenting with it. Yet despite this momentum, many organizations are struggling to translate AI into meaningful, scalable value.</p><p data-rte-preserve-empty="true" class="MsoNormal">And yet, across the industry, a consistent pattern is emerging: many organizations are approaching AI as a technology initiative.&nbsp; And that’s a mistake.</p><h2 data-rte-preserve-empty="true">The Technology Trap</h2><p data-rte-preserve-empty="true" class="MsoNormal">In many organizations, AI efforts follow a familiar path:</p><ul data-rte-list="default"><li><p data-rte-preserve-empty="true" class="MsoListParagraphCxSpFirst">evaluate tools</p></li><li><p data-rte-preserve-empty="true" class="MsoListParagraphCxSpMiddle">run pilot programs</p></li><li><p data-rte-preserve-empty="true" class="MsoListParagraphCxSpMiddle">test use cases</p></li><li><p data-rte-preserve-empty="true" class="MsoListParagraphCxSpLast">deploy isolated solutions</p></li></ul><p data-rte-preserve-empty="true" class="MsoNormal">This approach feels logical. It mirrors how many prior technologies were adopted.&nbsp;</p><p data-rte-preserve-empty="true" class="MsoNormal">But AI behaves fundamentally differently from prior technologies — in its speed of evolution, breadth of impact, non-deterministic nature, and rapid rate of obsolescence (among other reasons).&nbsp; AI does not simply support existing processes. It changes how work is performed, how decisions are made, and how value is created. It affects not just systems, but the structure and behavior of the institution itself.</p><p data-rte-preserve-empty="true" class="MsoNormal">Why does this distinction matter?&nbsp; Because organizations that focus primarily on tools and pilots often experience poor outcomes with AI:</p><ul data-rte-list="default"><li><p data-rte-preserve-empty="true" class="MsoListParagraphCxSpFirst">fragmented initiatives</p></li><li><p data-rte-preserve-empty="true" class="MsoListParagraphCxSpMiddle">higher error rates (e.g., hallucinations)</p></li><li><p data-rte-preserve-empty="true" class="MsoListParagraphCxSpMiddle">limited scalability</p></li><li><p data-rte-preserve-empty="true" class="MsoListParagraphCxSpMiddle">unclear value realization</p></li><li><p data-rte-preserve-empty="true" class="MsoListParagraphCxSpLast">growing complexity and risk</p></li></ul><p data-rte-preserve-empty="true" class="MsoNormal">The issue is not the technology.&nbsp; The issue is how the organization is aligned to use it.</p><h2 data-rte-preserve-empty="true">AI as a Capability Transformation Problem</h2><p data-rte-preserve-empty="true" class="MsoNormal">The organizations that are beginning to realize meaningful value from AI are not treating it as a technology layer.&nbsp; They are treating it as a capability transformation problem.&nbsp; This means asking a fundamentally different set of questions.</p><ul data-rte-list="default"><li><p data-rte-preserve-empty="true" class="MsoListParagraphCxSpFirst">What data foundations are required to support those decisions?</p></li><li><p data-rte-preserve-empty="true" class="MsoListParagraphCxSpMiddle">How should workflows be redesigned to incorporate AI outputs?</p></li><li><p data-rte-preserve-empty="true" class="MsoListParagraphCxSpMiddle">What governance and quality systems are needed to ensure reliability and trust?</p></li><li><p data-rte-preserve-empty="true" class="MsoListParagraphCxSpMiddle">What skills and roles are required to operate in this new environment?</p></li><li><p data-rte-preserve-empty="true" class="MsoListParagraphCxSpLast">How should decisions be made differently in an AI-enabled organization?</p></li></ul><p data-rte-preserve-empty="true" class="MsoNormal">AI does not succeed because a model works.&nbsp; AI succeeds when an institution is able to consistently and reliably integrate it into operations.</p><h2 data-rte-preserve-empty="true">A Pattern That Has Repeated Across the Industry</h2><p data-rte-preserve-empty="true" class="MsoNormal">Though AI is unique, healthcare and life sciences industry veterans may see some similarities with prior technology-driven transformations.</p><h3 data-rte-preserve-empty="true">Combinatorial Chemistry &amp; High Throughput Discovery</h3><p data-rte-preserve-empty="true" class="MsoNormal">One of the earliest transformations in modern drug development came with the rise of combinatorial chemistry and high-throughput screening.&nbsp; The promise was compelling: dramatically expand the number of compounds that could be generated and tested, accelerating the discovery of viable therapeutic candidates.</p><p data-rte-preserve-empty="true" class="MsoNormal">But success did not come simply from introducing new laboratory technologies.&nbsp; It required fundamental changes to how research organizations operated:</p><ul data-rte-list="default"><li><p data-rte-preserve-empty="true">new experimental designs and screening methodologies</p></li><li><p data-rte-preserve-empty="true">integration of automation technologies with laboratory workflows</p></li><li><p data-rte-preserve-empty="true">development of data systems capable of managing and interpreting vastly larger volumes of experimental results</p></li><li><p data-rte-preserve-empty="true">closer coordination between chemistry, biology, and computational teams</p></li></ul><p data-rte-preserve-empty="true" class="MsoNormal">Organizations that treated combinatorial chemistry as a discrete scientific innovation often struggled to realize its full value.</p><p data-rte-preserve-empty="true" class="MsoNormal">Even successful adopters did not fully realize the promised value. While combinatorial chemistry expanded the scale of compound generation and screening, it did not proportionally accelerate drug discovery.&nbsp; In many cases, the limiting factors were not the technologies themselves, but the surrounding systems required to translate scale into value. Biological understanding did not advance at the same rate as compound generation, data environments struggled to manage and interpret the resulting volume of information, and organizational models remained fragmented across chemistry, biology, and computational disciplines. As a result, organizations were able to generate more data and more candidate compounds, but not necessarily better outcomes.</p><p data-rte-preserve-empty="true" class="MsoNormal">The lesson was clear: scaling one component of the system without redesigning the broader operating model limits the impact of even the most powerful innovations.</p><h3 data-rte-preserve-empty="true">Digitizing Clinical Research</h3><p data-rte-preserve-empty="true" class="MsoNormal">In the early transition from paper-based trials to electronic data capture, success did not come from simply introducing new software.&nbsp; It required new processes; new data standards; new operating models; and cross-functional coordination across scientific, technical, and operational teams.&nbsp; The organizations that treated it as a system implementation struggled.&nbsp; Those that treated it as an institutional shift succeeded.</p><h3 data-rte-preserve-empty="true">Health Analytics</h3><p data-rte-preserve-empty="true" class="MsoNormal">As the healthcare industry adopted electronic medical records, their data became computable. &nbsp;So we saw the emergence of investments focused in data and analytics platforms supporting hospital operations, value-based care programs, and more.&nbsp; But analytics platforms alone did not produce impact.&nbsp; Real transformation required:</p><ul data-rte-list="default"><li><p data-rte-preserve-empty="true" class="MsoListParagraphCxSpFirst">building dedicated analytics capabilities that were adept at harnessing this data</p></li><li><p data-rte-preserve-empty="true" class="MsoListParagraphCxSpMiddle">aligning clinical and operational leadership on effective measurement strategies</p></li><li><p data-rte-preserve-empty="true" class="MsoListParagraphCxSpMiddle">adjusting care pathways, processes, and EMR functionality to leverage the insights</p></li><li><p data-rte-preserve-empty="true" class="MsoListParagraphCxSpMiddle">redefining decision-making processes based on the new insights available</p></li><li><p data-rte-preserve-empty="true" class="MsoListParagraphCxSpLast">establishing governance and prioritization frameworks protecting patients and the judgment of clinicians.</p></li></ul><p data-rte-preserve-empty="true" class="MsoNormal">Analytics became valuable not because it existed, but because it was institutionalized.</p><h3 data-rte-preserve-empty="true">The AI Era</h3><p data-rte-preserve-empty="true" class="MsoNormal">AI is now following the same pattern, but at far greater speed and scale.&nbsp; In areas such as clinical research, real-world evidence, regulatory processes, and healthcare delivery, AI is already reshaping how work is performed, how quickly it can be performed, and what level of precision is possible.&nbsp; In CROs, for example, AI is shifting value away from labor-based models toward AI-enabled capabilities, fundamentally changing how services are delivered and evaluated.</p><p data-rte-preserve-empty="true" class="MsoNormal">At the same time, organizations are struggling with competing pressures: speed vs. governance, innovation vs. quality, and adoption vs. control.&nbsp; Regulators are also signaling expectations that reinforce this shift — emphasizing data governance, lifecycle management, and multidisciplinary expertise as core requirements for AI-enabled systems.&nbsp; These are not technology challenges; they are institutional challenges.</p><h2 data-rte-preserve-empty="true">What It Actually Takes to Scale AI</h2><p data-rte-preserve-empty="true" class="MsoNormal">Organizations that succeed with AI tend to converge on a common set of capabilities.</p><ol data-rte-list="default"><li><p data-rte-preserve-empty="true" class="MsoNormal"><strong>Clear Strategic Alignment</strong>.&nbsp; AI initiatives are directly tied to measurable business and clinical outcomes, not isolated experimentation.</p></li><li><p data-rte-preserve-empty="true" class="MsoNormal"><strong>Data as a Managed Asset.</strong>&nbsp; Data is treated as a product — governed, curated, and designed for reuse — rather than an output of individual systems.</p></li><li><p data-rte-preserve-empty="true" class="MsoNormal"><strong>Redesigned Workflows.</strong>&nbsp; Processes are re-engineered to incorporate AI outputs into real decision-making, rather than layered on top of existing workflows.</p></li><li><p data-rte-preserve-empty="true" class="MsoNormal"><strong>Governance and Quality Systems.</strong>&nbsp; AI is managed within structured frameworks that ensure reliability, compliance, and trust.</p></li><li><p data-rte-preserve-empty="true" class="MsoNormal"><strong>Integrated Operating Models.</strong>&nbsp; Cross-functional teams align business, scientific, and technical perspectives to deliver and sustain capabilities.</p></li><li><p data-rte-preserve-empty="true" class="MsoNormal"><strong>Continuous Adaptation.&nbsp; </strong>Organizations are designed to evolve as technologies change, rather than relying on static roadmaps.</p></li></ol><h2 data-rte-preserve-empty="true">Moving Beyond the Pilot Phase</h2><p data-rte-preserve-empty="true" class="MsoNormal">Many organizations are still focused on identifying the “right” AI use cases.&nbsp; But that is not the limiting factor.&nbsp; The limiting factor is the ability to build the institutional capabilities required to scale those use cases.</p><p data-rte-preserve-empty="true" class="MsoNormal">Pilot programs are useful for learning.&nbsp; They are not sufficient for transformation.&nbsp; The organizations that move ahead will be those that shift their focus from “Where can we use AI?” to “What must we become to operate effectively in an AI-enabled world?”&nbsp; The real opportunity is not simply to do the same work faster.&nbsp; It is to transform the fundamentals:</p><ul data-rte-list="default"><li><p data-rte-preserve-empty="true" class="MsoListParagraphCxSpFirst">redesign clinical research</p></li><li><p data-rte-preserve-empty="true" class="MsoListParagraphCxSpMiddle">improve the precision of care</p></li><li><p data-rte-preserve-empty="true" class="MsoListParagraphCxSpMiddle">accelerate therapeutic development</p></li><li><p data-rte-preserve-empty="true" class="MsoListParagraphCxSpLast">fundamentally change how organizations operate and the associated costs</p></li></ul><p data-rte-preserve-empty="true" class="MsoNormal">The organizations that succeed will not be those that adopt AI tools the fastest.&nbsp; They will be those that build the capabilities required to make AI a core part of how they function.</p><p data-rte-preserve-empty="true" class="MsoNormal">AI is not a technology strategy issue. It is a transformation of how institutions create value.&nbsp; And the organizations that recognize this — and build accordingly — will define the next era of healthcare and life sciences.</p>]]></content:encoded><media:content height="1000" isDefault="true" medium="image" type="image/png" url="https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/1774379712035-GYS0TGAHIX3P9RAN6903/city-grid-chatgpt.png?format=1500w" width="1500"><media:title type="plain">AI Is Not a Technology Problem — It’s an Operating Model Transformation</media:title></media:content></item><item><title>Exploring the FDA and EMA Principles of Good AI Practice</title><dc:creator>Jason Burke</dc:creator><pubDate>Tue, 17 Mar 2026 17:06:46 +0000</pubDate><link>https://www.jasonburke.online/blog/exploring-the-fda-and-ema-principles-of-good-ai-practice</link><guid isPermaLink="false">674a65c6ba0c721888a63309:674a6885532ea52b2bdb097c:69c2c45e2b74012930e1cdf5</guid><description><![CDATA[Is your organization ready for AI regulation? Jason Burke unpacks the FDA 
and EMA's Good AI Practice guidance and the three capability gaps life 
sciences leaders must address now.]]></description><content:encoded><![CDATA[<p data-rte-preserve-empty="true">In January 2026, the FDA and EMA released a short but important document: “<a target="_blank" rel="noopener noreferrer nofollow" href="https://www.fda.gov/about-fda/artificial-intelligence-drug-development/guiding-principles-good-ai-practice-drug-development">Guiding Principles of Good AI Practice in Drug Development.</a>”</p><p data-rte-preserve-empty="true">At just two pages, the guidance is concise—but it sends a clear signal about how regulators expect artificial intelligence to be developed and governed across the drug development lifecycle.</p><p data-rte-preserve-empty="true">For an industry that is rapidly experimenting with AI, the document is both reassuring and clarifying. Let’s explore several of the key concepts they highlight.</p><h2 data-rte-preserve-empty="true">Natural Synergies with the Status Quo</h2><p data-rte-preserve-empty="true">For experienced industry leaders, several elements of the principles should be encouraging:</p><ol data-rte-list="default"><li><p data-rte-preserve-empty="true"><strong>Acceptance of the AI opportunity.</strong> If the FDA’s growing <a target="_blank" rel="noopener noreferrer nofollow" href="https://www.fda.gov/medical-devices/software-medical-device-samd/artificial-intelligence-enabled-medical-devices">portfolio of AI-enabled medical devices</a> and its <a target="_blank" rel="noopener noreferrer nofollow" href="https://www.fda.gov/regulatory-information/search-fda-guidance-documents/considerations-use-artificial-intelligence-support-regulatory-decision-making-drug-and-biological">2025 AI guidance document</a> were not enough to convince AI skeptics of regulatory readiness, this document reinforces the message. The agencies explicitly acknowledge that AI can help modernize the drug development paradigm, promoting innovation, reducing time-to-market, strengthening pharmacovigilance, and even reducing reliance on animal testing through improved prediction of toxicity and efficacy.</p></li><li><p data-rte-preserve-empty="true"><strong>Alignment with risk-based practices.</strong> The principles strongly reflect the risk-based ethos already embedded in regulatory science. This mirrors the operating philosophy used across areas such as monitoring, validation, quality systems, and clinical oversight. The industry already understands these frameworks, so the challenge ahead is extending them into AI-driven capabilities.</p></li><li><p data-rte-preserve-empty="true"><strong>Leveraging established validation approaches.</strong> One of the most common questions surrounding regulated AI is: how do you validate systems that are not fully deterministic? The answer suggested by the principles is reassuringly familiar. The agencies emphasize practices already central to regulated technology environments: robust design, data governance, engineering best practices, documentation, and quality management systems among other topics. These expectations are fully consistent with how technology-driven solutions are already developed and validated in regulated environments.</p></li></ol><p data-rte-preserve-empty="true">This is all good news, as it further confirms what many industry insiders have been expecting.</p><h2 data-rte-preserve-empty="true">The Role of Standards</h2><p data-rte-preserve-empty="true">Despite its brevity, the document references “standards” five times. Historically, US and EU regulators have relied heavily on standards to streamline their complex and safety-critical activities. Standards create a shared language that simplifies compliance and inspection for both sponsors and regulators. The challenge is that standards require stability, and stability is not a near-term feature of today’s rapidly-evolving AI landscape. </p><p data-rte-preserve-empty="true">Both agencies are encouraging industry collaboration to further develop and implement these principles. Could we see the formation of an organization similar to CDISC, but focused on AI standards in drug development? Perhaps. </p><p data-rte-preserve-empty="true">Having been involved in the formation of CDISC, I remember firsthand how complex that effort was. I would also argue that CDISC’s data domains — mainly limited to information structured in research protocols and statistical analysis plans — are more constrained than today’s AI models and use cases. Practical industry applications of AI today routinely feature electronic medical records, genomics and molecular data, patient registries, claims and prescription data, lab results, and other real-world data (RWD) and real-world evidence (RWE). Many of AI’s most transformative applications will depend on combining these heterogeneous data environments, which introduces significant challenges for standardization.</p><h2 data-rte-preserve-empty="true">Industry Readiness</h2><p data-rte-preserve-empty="true">For many industry leaders, the biggest question raised by the document is simple: is our organization actually ready? </p><p data-rte-preserve-empty="true">Many life sciences companies are still pursuing AI opportunistically through pilot projects, experimentation, and point-solution tools. While these activities can generate useful experiences, they do not inherently cultivate the disciplined operating environments regulators are describing. They also don’t protect the organization from the risks in AI permeating every system and vendor within the enterprise.</p><p data-rte-preserve-empty="true">In practice, three capability gaps tend to emerge:</p><p data-rte-preserve-empty="true"><strong>Lack of an Organizing Framework.</strong> The guidance repeatedly references the importance of “context of use”, including role, scope, model risk, and quality procedures. Establishing this clarity requires an organizational framework for AI governance aligned to the company’s quality system. Many organizations have not developed such a framework and lack strategies for ensuring adherence to it.</p><p data-rte-preserve-empty="true"><strong>Compliance &amp; Inspection Evidence.</strong> The principles closely mirror areas regulators traditionally examine during inspections, including data provenance, process adherence, audit trails, data protections, and documentation. Ironically, these are often the areas that receive less attention during rapid technological exploration. In AI environments, weaknesses in these areas do more than create regulatory risk — they frequently undermine the quality, reliability, and performance of the AI solutions themselves.</p><p data-rte-preserve-empty="true"><strong>AI-specific Processes and Expertise.</strong> The guidance emphasizes the need for multidisciplinary expertise that spans both AI technology and the scientific or business context in which it is used. Many organizations have yet to create formal mechanisms that support this type of AI-informed collaboration. Equally important, AI introduces new lifecycle management challenges, including model monitoring, drift detection, retraining / tuning, migration, and evolving cybersecurity risks. In practice, many companies simply do not have sufficient internal AI expertise to fully address AI competencies that deliver engineering best practices, interpretability, explainability, performance controls, transparency, generalizability, and robustness.</p><p data-rte-preserve-empty="true">As AI adoption accelerates, these gaps should be a call to action. </p><h2 data-rte-preserve-empty="true">Leaning into Next</h2><p data-rte-preserve-empty="true">The encouraging news is that these challenges are not insurmountable. Proven governance, quality, and technology management approaches can address them—when applied deliberately and supported by the right expertise.</p><p data-rte-preserve-empty="true">For leadership teams, several strategic questions become increasingly important:</p><ul data-rte-list="default"><li><p data-rte-preserve-empty="true">Who owns our AI readiness and what are the specific goals and outcomes we are pursuing? </p></li><li><p data-rte-preserve-empty="true">How are we orchestrating change management, cross-functional alignment, and standards development around AI adoption?</p></li><li><p data-rte-preserve-empty="true">How are we prioritizing improvements to our quality system, data environments, and development practices so that concepts like “context of use” are reflected in daily operations?</p></li><li><p data-rte-preserve-empty="true">Where should we supplement internal teams with external AI expertise to ensure regulatory and safety considerations are properly managed? </p></li></ul><p data-rte-preserve-empty="true">The answers to these questions will not emerge overnight. And AI capabilities will continue to evolve rapidly for the foreseeable future. What life sciences organizations need is not a one-time solution, but a sustainable way to bring structure and discipline to what often still feels like the “Wild West” of AI.</p><p data-rte-preserve-empty="true">The organizations that succeed will be those that balance innovation with governance—moving quickly while building the operational discipline regulators clearly expect.</p><p data-rte-preserve-empty="true">How prepared is your organization for that shift?</p>]]></content:encoded><media:content height="841" isDefault="true" medium="image" type="image/jpeg" url="https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/1774372136379-QPNJJY0N8R98AX05HCQP/digital-pill-image.jpg?format=1500w" width="1500"><media:title type="plain">Exploring the FDA and EMA Principles of Good AI Practice</media:title></media:content></item><item><title>Why AI Feels Like the Wild West</title><dc:creator>Jason Burke</dc:creator><pubDate>Tue, 10 Mar 2026 17:01:27 +0000</pubDate><link>https://www.jasonburke.online/blog/why-ai-feels-like-the-wild-west</link><guid isPermaLink="false">674a65c6ba0c721888a63309:674a6885532ea52b2bdb097c:69c2c2dc1a0de941e16cffd3</guid><description><![CDATA[AI in life sciences feels chaotic — but it doesn't have to. Explore the 7 
key tensions facing industry leaders and the principles that separate AI 
winners from the rest.]]></description><content:encoded><![CDATA[<h2 data-rte-preserve-empty="true">The Showdowns Defining AI in Life Sciences Today</h2><p data-rte-preserve-empty="true">The rapid expansion of the American West created a landscape defined by opportunity, competition, and uncertain rules. Today, many life sciences leaders may feel they are navigating a similar frontier with artificial intelligence.</p><p data-rte-preserve-empty="true">The AI territories for industry transformation look both extensive and attractive: drug design, trial simulation, data cleaning, regulatory writing, quality monitoring, employee self-service, and many more paint a vibrant picture of AI-enabled businesses that deliver better therapies faster. But the day-to-day experience can sometimes feel like a showdown of competing concerns:</p><ol data-rte-list="default"><li><p data-rte-preserve-empty="true"><strong>Velocity vs. Planning.</strong> Frontier AI companies are releasing product updates at an unprecedented pace – measured in months, not years. Combined with the continuous influx of new AI-enabled players in the market, corporate AI roadmaps become outdated almost as quickly as they are authored.</p></li><li><p data-rte-preserve-empty="true"><strong>Use vs. Proliferation.</strong> Industry leaders want to encourage the advancements and efficiencies that AI can bring (e.g., faster recruitment, streamlined study operations, cleaner clinical data). But AI growth brings practical challenges: proliferating data copies, rising compute costs, tool sprawl, and increasing compliance risk.</p></li><li><p data-rte-preserve-empty="true"><strong>Agility vs. Governance.</strong> From the board room to the break room, leaders and workers are embracing AI — from tools that summarize clinical trial documents to copilots that draft regulatory submissions. As organizations struggle to make decisions and implement new solutions, research shows a large portion of employee adoption is “shadow AI” use of ungoverned and unsecured products.</p></li><li><p data-rte-preserve-empty="true"><strong>Market Enthusiasm vs. Value Creation.</strong> The market noise around AI is deafening, and expectations are frequently misaligned to reality. Leaders struggle to balance the imperative to move with a disciplined strategy that focuses on the most attractive and measurable value creating opportunities.</p></li><li><p data-rte-preserve-empty="true"><strong>Duplication vs. Standardization.</strong> Practically every IT vendor in every organizational unit is bringing AI to market. Leaders responsible for quality deliverables, compliant operations, and controlled costs are grappling with a deluge of independent models offering similar capabilities, differing reliability, and growing costs.</p></li><li><p data-rte-preserve-empty="true"><strong>Build vs. Buy Decisions.</strong> Provisioning AI capabilities is not a simple “build vs. buy” equation, and thanks to AI, tailored solutions don’t necessarily require manual coding. Though AI tools can be purchased “off the shelf,” more impactful AI solutions that deliver higher value for any organization are often some combination of both.</p></li><li><p data-rte-preserve-empty="true"><strong>Change vs. Quality.</strong> Business transformation with AI requires workforce education, process re-engineering, and evolving roles. In the absence of disciplined programs focused on effective transitions, the corresponding AI capabilities frequently face problems in accuracy, scalability, compliance, trust, adoption, and ROI.</p></li></ol><h2 data-rte-preserve-empty="true">How the West is Won</h2><p data-rte-preserve-empty="true">The AI frontier may feel chaotic, but organizations that succeed are not improvising. They operate with a clear set of principles that guide technology, governance, and change.</p><ul data-rte-list="default"><li><p data-rte-preserve-empty="true"><strong>One Program, not Many Projects.</strong> Scalable AI capabilities with measurable value creation do not emerge naturally from grass-roots efforts. As opposed to waiting to uncover small-scale wins, create an orchestration center that can drive a more managed approach across the business. It is faster and cheaper to create the right investments once.</p></li><li><p data-rte-preserve-empty="true"><strong>Strategy, not Roadmaps.</strong> The AI landscape will not stabilize any time soon. Instead of trying to maintain static adoption roadmaps that are always out of sync with the market, pivot to focusing on clear strategies, technical principles, and change-ready blueprints that guide the ongoing selection and migration of AI capabilities over time as the business and technology evolve.</p></li><li><p data-rte-preserve-empty="true"><strong>Data, not Tech.</strong> Most risks in AI projects have nothing to do with the AI itself. Experience shows that most organizations do not have data assets designed to perform well with AI (e.g., semantically consistent research data; managed master data; data architectures that effectively blend structured and unstructured data across scientific, operational, and financial domains). Instead of evaluating the next great AI tool, direct teams to carefully assess the quality, consistency, management, and trustworthiness of the organization’s key data that would be used by any AI model selected.</p></li><li><p data-rte-preserve-empty="true"><strong>Quality System, not Documentation.</strong> The organization’s obligation to cultivating and using high-quality information systems – whether regulated or not – grows with AI. When systems are not deterministic, technology is changing, and processes are being re-engineered, it is critical to address quality as a comprehensive framework guiding effective, safe, and compliant business operations. </p></li><li><p data-rte-preserve-empty="true"><strong>Enablement, not Deployment.</strong> To unlock sustainable value creation with AI, adoption is usually a transformational journey. The goal is not to give employees more software. Focus programs on enabling employees to perform different processes, empowering them to co-create the future, leveraging new skills, and freeing capacity from manual labor that machines can undertake.</p></li></ul><p data-rte-preserve-empty="true">The AI frontier in life sciences will not normalize anytime soon. But history suggests that periods of technological expansion eventually reward the organizations that combine ambition with discipline. In the AI era, the winners will not simply adopt new tools faster — they will build the strategies, data foundations, and operating models that allow those tools to create lasting value.</p>]]></content:encoded><media:content height="844" isDefault="true" medium="image" type="image/jpeg" url="https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/1774371716443-GHE8RES94ZGJ5Y7IYP70/lab-image.jpg?format=1500w" width="1500"><media:title type="plain">Why AI Feels Like the Wild West</media:title></media:content></item><item><title>CROs in the Age of AI</title><dc:creator>Jason Burke</dc:creator><pubDate>Tue, 03 Mar 2026 17:27:00 +0000</pubDate><link>https://www.jasonburke.online/blog/cros-in-the-ai-age</link><guid isPermaLink="false">674a65c6ba0c721888a63309:674a6885532ea52b2bdb097c:69c2bb546d7c4013b8502cdd</guid><description><![CDATA[Discover how AI is transforming the $100B CRO industry — from business 
model shifts to data strategy — and what clinical research leaders must do 
to stay competitive.]]></description><content:encoded><![CDATA[<p data-rte-preserve-empty="true">How do people-driven businesses adapt to the impact of artificial intelligence (AI)?</p><p data-rte-preserve-empty="true">That’s the top-of-mind question for leaders across many service sectors. Compared to prior emerging technologies, AI is somewhat unique: in addition to its exceptionally high velocity and broad applications, the disruption is felt heavily among skilled knowledge workers. Whereas technology innovations often empower and amplify this workforce, AI also commoditizes the delivery of knowledge work products and processes. In life sciences, contract research organizations (CROs) are one of the leading intersections of this workforce / AI transformation.</p><p data-rte-preserve-empty="true"></p><p data-rte-preserve-empty="true">The modern CRO industry has evolved over the past 40 years to become a diverse $100B ecosystem of service providers responsible for &gt;70% of all clinical research conducted in the world. But the fundamental value propositions historically offered by CROs – operational efficiencies, scalable resource models, cost arbitrage, shared IT investment, and access to scientific expertise – are being re-characterized by the rapid infusion of AI. As these economics shift, CROs face an imperative to adapt both their business models and operations to remain competitive and valuable business partners.</p><h2 data-rte-preserve-empty="true">Pathways to Smarter Clinical Research</h2><p data-rte-preserve-empty="true">Though a full-scale disruption of the CRO industry is unlikely, a transition in service delivery approaches and associated value creation – from labor-oriented models towards diversification into AI-enabled products and services – is already underway. Market viability of AI solutions has already been proven in areas such as:</p><ul data-rte-list="default"><li><p data-rte-preserve-empty="true">Site selection and start-up</p></li><li><p data-rte-preserve-empty="true">Patient recruitment and retention</p></li><li><p data-rte-preserve-empty="true">Adaptive trial designs &amp; synthetic control arms</p></li><li><p data-rte-preserve-empty="true">Clinical / RWE data management and quality</p></li><li><p data-rte-preserve-empty="true">Adverse event detection and management</p></li><li><p data-rte-preserve-empty="true">Imaging, wearables, and devices</p></li><li><p data-rte-preserve-empty="true">Clinical monitoring and protocol deviation detection</p></li><li><p data-rte-preserve-empty="true">Document automation solutions</p></li><li><p data-rte-preserve-empty="true">Programming and TLF automation</p></li><li><p data-rte-preserve-empty="true">Real world evidence analyses and NLP</p></li></ul><p data-rte-preserve-empty="true">To date, the AI investment strategy of CROs in these areas has been stratified by market tier.</p><ul data-rte-list="default"><li><p data-rte-preserve-empty="true"><strong>Large CROs</strong> are building proprietary platforms, making notable investments alongside global technology players and partners to deploy large-scale data assets and extensible agent platforms. These complex systems require sizable capital investments and expertise, and while they offer compelling capabilities, they will also require sustained engineering as the underlying technologies continue to evolve.</p></li><li><p data-rte-preserve-empty="true"><strong>Small-to-medium CROs</strong>, lacking the capital of their larger peers, are pursuing 3rd party vendors to provision off-the-shelf AI capabilities, though such tactics introduce a wide array of risks (e.g., vendor dependency, parity with competitors, operational misalignment, cost duplication, data proliferation) that can potentially undermine long-term value creation.</p></li></ul><p data-rte-preserve-empty="true">For most of the CRO industry, a balanced approach to build-vs.-buy is likely the best path forward. The introduction of value-creating AI solutions does not need to be capital intensive – on the contrary, most technical capabilities can be provisioned at commodity price points. By blending affordable off-the-shelf software with tailored configurations of supportable AI, CROs can unlock the transformative value of AI without the need for lengthy, complex system engineering projects and high operating expenses.</p><h2 data-rte-preserve-empty="true">Business Model and Market Dynamics Implications</h2><p data-rte-preserve-empty="true">Given the capabilities that AI offers, performance expectations for both sponsors and CROs will change. Though geographic reach, operational scale, and clinical expertise will still matter, sponsors will evaluate CROs on their ability to deliver AI-derived advantages such as faster performance (e.g., recruitment, data review, database lock), more efficient operations (e.g., study monitoring, programming), higher quality (e.g., lower error rates, automated compliance and quality monitoring), and more precise results (e.g., modeling, targeting, diagnostics, segmentation). As such, leadership teams and their boards need to be prepared to proactively address strategic questions such as:</p><ul data-rte-list="default"><li><p data-rte-preserve-empty="true">Business model: Every CRO needs a strongly aligned strategy between their relationship model (FSO, FSP, hybrid), their differentiated capabilities (therapeutic expertise, accessible network, data, technology, etc.), and their AI-driven value delivery capabilities (speed, quality, cost, risk, precision, etc.).  <em>What makes your CRO a truly unique value creation driver in an AI-enabled market, and how will you best enable and monetize it?</em></p></li><li><p data-rte-preserve-empty="true">Investor attractiveness: In an sector historically characterized by high investor appetites, CROs that do not demonstrate strong AI and data competencies may be seen as less competitive and garner lower valuations.  <em>How will you align to investor interests to maximize future capitalization cycles?</em></p></li><li><p data-rte-preserve-empty="true">Margins: Some CROs may require an increase in highly skilled labor (e.g., scientific / technical expertise delivering novel research designs involving synthetic control arms, digital twins, and trial simulations). <em>How will you protect and improve margins with higher labor costs, especially if lower-cost resources are displaced?</em></p></li><li><p data-rte-preserve-empty="true">Continuous learning and change: For the foreseeable future, the rate of AI-driven advancements and change will remain high, presenting the CRO industry with a continuous stream of opportunities. <em>How will you stay agile and embrace new AI opportunities effectively while avoiding unnecessary expenses and technical debt?</em></p></li><li><p data-rte-preserve-empty="true">Consolidation: Given the market velocity, both CROs and AI companies seeking to accelerate their growth will see inorganic options as a competitive imperative. Consolidations will change both CRO competitive dynamics and vendor technology strategies. <em>How will you monitor and adapt to these landscape changes, and which ones would trigger adjustments to your strategy?</em></p></li></ul><h2 data-rte-preserve-empty="true">Practical Considerations for AI Enablement of CROs</h2><p data-rte-preserve-empty="true">When considering internal investments, “AI everywhere” will not create high performance. Every CRO needs an AI strategy that prioritizes the fewest, most important operational and competitive capabilities that will drive the highest value creation. The AI strategy should be deeply connected to the corporate strategy, focusing on performance metrics tied to explicit organizational goals. Examples include margin expansion, utilization improvement, avoiding unnecessary human capital growth, study enrollment optimization measures, improved turnaround times, improved quality, and higher win rates. For most CROs, that strategy will need to address the follow 7 areas.</p><ol data-rte-list="default"><li><p data-rte-preserve-empty="true"><strong>Service line transformation.</strong> Existing service lines will evolve – and new service lines will emerge – to accommodate opportunities that AI introduces. Existing service lines will unlock efficiencies in core business processes such as recruitment, data review, scientific writing, biostatistical programming, and billing. Newer services reflect growing demand for bespoke client-driven AI applications – novel study modeling, disease modeling, and RWE data curation such as patient registries, for example. One key to unlocking the strategic value of service line transformation is laser-like clarity around a firm’s most compelling business, scientific, and technical differentiators that can be translated into AI assets.</p></li><li><p data-rte-preserve-empty="true"><strong>Relationship models.</strong> Sponsor relationships with CROs can take the form of any combination of T&amp;M projects, Full-Service Outsourcing (FSO), and Functional Service Provider (FSP) agreements. Hybrid approaches and FSP relationships tend to be more cost effective for clients, though smaller organizations such as some biotechs may require FSO providers’ infrastructure. Though AI can amplify CRO margins of all three options, it also places downward pricing pressure on labor-based contracting even under milestone-based payment models. Outcomes- or performance-based contracting can provide margin relief by leveraging AI to drive value creation provided adopting organizations ensure strong management of the contract terms.</p></li><li><p data-rte-preserve-empty="true"><strong>Data sharing readiness.</strong> Ensuring that client agreements provide data reusability terms to support the CRO’s AI-enablement goals is now vital. For some, changing contract terms will seem commercially risky. But the alternative is much riskier – an organization based heavily on manual labor competing in an industry at least partially automated by AI. Numerous precedents exist for safely sharing anonymized data to support research advancements. And not all data is equally sensitive; operational, quality, and performance related data offer a wealth of innovation opportunities without introducing scientific IP, patent, or patient data risks. Accommodations may be needed for individual customers that may be unwilling, though offering clear incentives for sharing may be sufficient to produce change.</p></li><li><p data-rte-preserve-empty="true"><strong>Data asset development.</strong> Though market excitement is often directed at advancements in AI methods, success and value creation with AI is most associated with data. In some areas such as clinical development, the historical approach to data management has been largely study- or program-based; cultivating reusable data assets has not been a high priority. A CRO may have a strategy that focuses on any combination of preclinical, clinical, financial, commercial, and operational data. Regardless of the data mix, competencies such as data product engineering, quality, curation, governance, pedigree, ownership, and performance-aligned storage are needed to ensure AI success. And controlling data proliferation – the uncontrolled spread of data copies across AI tools and storage locations – is critical for protecting information security.</p></li><li><p data-rte-preserve-empty="true"><strong>Real-world data partnerships.</strong> Many AI innovation priorities for CROs will not be accessible without better access – either directly or through partners – to real-world data. AI-enabled patient recruitment, for example, is most effective when connected to patient medical records. Advanced work with precision diagnostics and therapeutics involves leveraging richer data formats – sequence data, imaging data, waveforms, streaming device data, and others. CROs do not need to in-source all of these data capabilities, but they do need consistent, scalable channels for provisioning them so they can develop reusable IP. Therapeutic specialization offers unique value accelerator; organizations establishing unique data niches within specific therapeutic areas or modalities gain additional commercial and computational leverage.</p></li><li><p data-rte-preserve-empty="true"><strong>Enterprise architecture.</strong> For AI to function and scale properly, the IT strategy, blueprint, and underlying systems need to be intentionally designed, aligned, and managed. Every IT vendor in the CRO’s enterprise is pitching AI capabilities. A clear plan for governance and rationalized use is required for scale and cost management. CROs need standardized designs and solutions for addressing common AI scenarios such as access to external models, delivery of tailored AI solutions, authentication, and data controls. As AI use cases increasing focus on workflow automation, interoperability within and across enterprise systems becomes an imperative. As many industry applications were never engineered as enterprise-grade systems, organizations may need some tailored workarounds to accomplish their integration goals. And given that the AI technology landscape will continue to evolve rapidly over the next five years, every CRO needs to plan for continuous change – leverage what is available but avoid compounding technical debt.</p></li><li><p data-rte-preserve-empty="true"><strong>Analysis.</strong> In its traditional form, biostatistics may now be a less viable independent service line. While still core to research execution and interpretation, traditional bioanalytical methods (e.g., descriptive and inferential statistics, Bayesian approaches) represent only a fraction of the computational techniques now being applied in supporting drug discovery and development. In addition, the implementation of those traditional methods (e.g., programming, testing, deliverable creation) can be delivered by modern AI models, creating a barrier to high-margin service monetization. However, the human expertise in the appropriate selection, use, and interpretation of AI and bioanalytical methods is not a commodity, creating tremendous opportunity for leveling up the contribution margin of analytical talent.</p></li></ol><p data-rte-preserve-empty="true">From a leadership perspective, perhaps the biggest consideration to AI adoption in life sciences – at least in terms of speed and impact – is the industry environment itself. Research variability, scientific complexity, regulatory obligations (both real and perceived), and justifiable concerns over patient protections have often created conservative, less agile cultures. No one wants to be at the source of an audit failure, a delayed study close, or loss of IP. At the same time, many success stories now exist, regulators are actively participating in the journey, and life sciences has successfully adopted many emerging technologies over the past three decades. Leaders need to acknowledge the legitimate nature of these concerns while also delivering confidence-inspiring plans for navigating into the future.</p><h2 data-rte-preserve-empty="true">Coming of Age</h2><p data-rte-preserve-empty="true">CROs that move now to embrace the age of AI will be tomorrow’s market leaders and high-value investment targets. But AI adoption is a learning journey for every organization, including CROs. For leaders looking to accelerate their AI progression, three tactics show up repeated as best practices among high-performing organizations:</p><p data-rte-preserve-empty="true"><strong>1. Conduct a comprehensive, realistic assessment of your organization’s readiness.</strong> Given the issues presented here, it is not surprising that most CROs are not ready to fully embrace AI. Leaders need to be prepared to critically evaluate the maturity of processes and competencies related to:</p><ul data-rte-list="default"><li><p data-rte-preserve-empty="true">Business process standardization and quality control</p></li><li><p data-rte-preserve-empty="true">Data and information storage, control, quality management, access, and use agreements</p></li><li><p data-rte-preserve-empty="true">Solution development lifecycles, quality frameworks, and auditability</p></li><li><p data-rte-preserve-empty="true">IT governance, including vendor assessments and data use terms</p></li></ul><p data-rte-preserve-empty="true"><strong>2. Develop specific, measurable value creation goals aligned to your corporate strategy and growth opportunities.</strong>Avoid searching for perfect AI applications. Instead, look for current business problems where AI can help. Prioritize measurable constraints such as situations where you:</p><ul data-rte-list="default"><li><p data-rte-preserve-empty="true">Lose margin due to manual work</p></li><li><p data-rte-preserve-empty="true">Struggle with people capacity</p></li><li><p data-rte-preserve-empty="true">Have people allocated below their skill level • See unmet service quality, capability, or speed</p></li><li><p data-rte-preserve-empty="true">Find manual work producing delays</p></li><li><p data-rte-preserve-empty="true">Cannot easily access knowledge resources</p></li></ul><p data-rte-preserve-empty="true"><strong>3. Launch, don’t pilot, AI solution programs to gain the right expertise.</strong> Pilot programs are useful when testing whether a technology works, but they are much less useful when pursuing value creation goals, process re-engineering, or business transformation. The core question is not whether the technology is a good fit for the business today; the core question is about what the business can look like tomorrow. We know the technology works; the goal is creating a more attractive business with it. And the goal is within reach.</p>]]></content:encoded><media:content height="841" isDefault="true" medium="image" type="image/jpeg" url="https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/1774370424899-UFE00NT5GTEDUFHTJ86E/network-city-image.jpg?format=1500w" width="1500"><media:title type="plain">CROs in the Age of AI</media:title></media:content></item><item><title>Technology Design in the Age of AI</title><category>AI</category><dc:creator>Jason Burke</dc:creator><pubDate>Tue, 15 Jul 2025 18:37:00 +0000</pubDate><link>https://www.jasonburke.online/blog/ai-technology-design</link><guid isPermaLink="false">674a65c6ba0c721888a63309:674a6885532ea52b2bdb097c:688bb7de0094ff321b2af47a</guid><description><![CDATA[AI is disrupting over 75 years of computer-related design. The changes 
reflect an amazing inflection point that both business and technology 
leaders need to reflect in their strategies.]]></description><content:encoded><![CDATA[<p class="">I’m passionate about design. I see beauty and genius in discovering novel structures and patterns that are functional and elegant in their simplicity. So I’m surprised more people are not talking about AI disrupting over 75 years of computer-related design. To me, it seems like an amazing inflection point.</p><h3><strong>Current Design Context</strong></h3><p class="">The term “computer” was originally assigned to human beings – in particular, individuals that compute things for others. Acknowledging the routine work associated with computing answers, Charles Babbage developed the concept of a mechanical programmable computer (some people may recall that GameStop’s original name was Babbage’s). Electromechanical computers followed in the 1930s-1940s, with <a href="https://en.wikipedia.org/wiki/Colossus_computer"><strong>Colossus</strong></a> and <a href="https://en.wikipedia.org/wiki/ENIAC"><strong>ENIAC</strong></a> becoming the first electronic digital programmable computers. Through the rise of transistors, integrated circuits, and today’s most powerful processors and architectures, one design principle has held constant: humans would be the main users of this technology.</p><p class="">But AI is challenging long-standing hardware and software architectural and design concepts. I’ve compiled a list of the biggest design-related shifts I see emerging from the evolving capabilities in AI.</p><h3><strong>Computers as First-Class Users</strong></h3><p class="">To many people, the most fundamental and obvious shift broadens the intended developer and user community from humans to include machines. Though information systems have participated in workflows for decades, AI-empowered systems feel like a different design metaphor.</p><ul data-rte-list="default"><li><p class=""><strong>AI is a primary developer and user.</strong> Software will no longer primarily be built just for people to click and type. Instead, AI programs themselves are becoming priority users of digital systems, fundamentally changing how everything is designed from the ground up. Systems will become optimized for autonomous AI interactions, and new software applications will be built by software.</p></li><li><p class=""><strong>AI becomes inseparable from software engineering.</strong> The process of building software is being transformed by AI. AI tools will assist developers in generating code, creating comprehensive test suites, discovering obscure edge cases, and predicting potential build failures.<strong> </strong>The role of human software developers will evolve from writing every line of code to becoming guides and supervisors for AI tools, focusing on designing the overall system, defining complex problems, and overseeing AI-generated solutions. This of course changes the design of integrated development environments (IDEs).</p></li><li><p class=""><strong>Autonomous AI agents perform routine work.</strong> Future software will be powered by intelligent programs, or AI agents, that can observe, reason, and act independently. These agents can plan their own steps and execute complex tasks without needing constant human oversight. Designs will need to reflect this new operating autonomy. And AI will move into IT operations, leading to self-healing systems. AIOps platforms will use machine learning to monitor software performance in real-time, predict potential issues, guard against cybersecurity threats, initiate maintenance, and even automatically resolve problems without human intervention.</p></li><li><p class=""><strong>Data gets designed and aggregated for machines:</strong> The way information is stored and organized is changing. Data architectures will be built considering the need for AI to process and interpret data efficiently, rather than just for humans to read or analyze. Vector databases and other emerging models allow AI to find things based on their meaning and similarity, not just exact keywords.</p></li></ul><h3><strong>Software Value Flows from Data</strong></h3><p class="">For most of the digital computing age, software value derived from the code – the algorithms and associated functionality they deliver. But as code becomes an automated commodity, AI considerably alters that value equation:</p><ul data-rte-list="default"><li><p class=""><strong>Data is a differentiator.</strong> General-purpose AI models are rapidly becoming a commodity. But specialized models – trained and tuned to specific tasks using highly curated data – are not a commodity. And that means organizations that invest in these data assets – and the software solutions that generate that data – will enjoy more upside.</p></li><li><p class=""><strong>Data timeliness gains prominence</strong>. AI models will increasingly rely on diverse methods of perceiving the world around them, with data flowing continuously and instantly throughout software systems. Robotics and sensors will broaden the focus from camera and microphone inputs to include multispectral vision, spatial audio, haptics, balance, position, velocity, olfaction, bioelectric, chemical, energy, gravimetric, telemetry, and even internal state.</p></li></ul><h3><strong>Software Becomes More Human &amp; Dynamic</strong></h3><p class="">There was a time when technologists needed a lot of user manuals. Those days are coming to an end as software archetypes reflect a bias towards human interactions.</p><ul data-rte-list="default"><li><p class=""><strong>Conversational interactions drive information delivery.</strong> Instead of navigating menus or learning complex commands, users will increasingly talk or type naturally to software. Interacting with systems becomes less about functional mastery (i.e., knowing the software) and more about subject mastery (i.e., understanding what needs to be done).</p></li><li><p class=""><strong>Speech becomes a preferred I/O model.</strong> Because application functionality will be delivered through conversational exchanges, it will seem more natural and expedient to simply talk to some AI models and agents. This voice mode option means that AI interactions will occur in times and places where typing and screens would have limited convenient use of earlier AI systems.</p></li><li><p class=""><strong>Intent-driven software will displace many linear applications.</strong> The software of the future won't just follow step-by-step instructions. It will be designed to understand high-level goals and intentions, then figure out the best way to achieve them. This means users will tell the system what needs to be done and why, but not how to do it.</p></li><li><p class=""><strong>Application functionality will be built on-the-fly</strong>. Historically, application functionality is designed, built, and tested as part of a product R&amp;D cycle. Though that will continue, functionality will also be dynamically coded on demand by AI models. The models will discern the user’s request, build a tiny program to deliver it, execute the program, and return the results.</p></li><li><p class=""><strong>Contextualized interfaces replace one-size-fits-all user experiences.</strong>Software user interfaces have gradually been getting smarter for decades, shifting from a static set of menus to context-aware displays and functionality. AI will push that trend even further with interfaces that are generated dynamically according to individual needs, preferences, and the activity currently underway.</p></li></ul><h3><strong>Integration As A Primary Design Criterion</strong></h3><p class="">Most organizations will see AI as a means to automate workflow across systems. As such, the ability to connect differing systems and data together moves from a “feature” to a critical design element that drives value delivery from AI models.</p><ul data-rte-list="default"><li><p class=""><strong>Integrations are built for and with AI.</strong> The connections between different software systems (e.g., APIs and other data interchange functionality, protocols, and standards) will be designed specifically for intelligent AI agents. These mechanisms will allow AI programs to communicate more effectively, maintain context across interactions, and use tools more flexibly. To ensure different AI models and systems can work together seamlessly, new industry standards and protocols will emerge that define how AI agents communicate capabilities, exchange context, and integrate across various platforms.</p></li><li><p class=""><strong>Retrieval-oriented AI designs will be everywhere</strong>. Since AI models need access to enterprise data repositories for processing (e.g., RAG), solution architectures that explicitly incorporate retrieval capabilities will be a common practice. Enterprise systems, AI agents, MCP servers, and workflows will all incorporate this design requirement.</p></li><li><p class=""><strong>Modular AI microservices deliver extensible functionality.</strong> Instead of one large program, software will be broken down into many small, independent AI capabilities that operate as separate services. This microservices approach allows each AI function to be developed, scaled, and updated independently, making systems more flexible and robust.</p></li></ul><h3><strong>Computing Paradigms Expand</strong></h3><p class="">From its origins as room-sized machines through the handheld computing revolution and wearables, information technologies have become more pervasive and accessible. AI will further amplify this perfusion of technology.</p><ul data-rte-list="default"><li><p class=""><strong>Hardware options will further diversify.</strong> Because many AI capabilities will be cloud-resident, a broader array of hardware form factors will be able to use them. Typical constraints – screen and keyboard access, for example – become less important than connectivity to networks and sensors that can feed data to models and rapidly interpret responses.</p></li><li><p class=""><strong>AI operating systems will emerge.</strong> Historically, operating systems have existed to enable user access and control of hardware. With the shift in focus from human to machine users, operating systems will evolve from basic hardware control systems and user-facing features to agent and task control systems designed to support autonomous applications.</p></li><li><p class=""><strong>Many AI deployments will be services, not servers.</strong> Many AI tasks, especially those that run occasionally or have unpredictable demand, will use serverless computing. This means the cloud provider automatically manages the underlying infrastructure, and users pay for compute time. This approach changes the economics of many corporate IT strategies.</p></li><li><p class=""><strong>AI deployments will also be device-resident</strong>. As AI methods evolve and mature, many AI models will be small, capable of being deployed within individual devices. Localized rapid processing, especially in fields such as computer vision and language processing, help ensure low latency responses and resilience to network disruptions.</p></li><li><p class=""><strong>Event-driven designs drive scalability.</strong> Software components will communicate by reacting to "events"—significant changes in the system's state. This event-driven architecture enables AI agents to respond instantly to business events, triggering autonomous actions and making the entire system more responsive and scalable.</p></li></ul><h3><strong>Safety as a Design Priority</strong></h3><p class="">Other sectors – automotive, children’s toys, and more – have long embraced safety as a critical element of product design. Beyond industry-specific use cases, historical software designs have not needed to consistently embrace a safety mandate, but it will be non-negotiable for many AI technologies, whether stipulated in regulations or not.</p><ul data-rte-list="default"><li><p class=""><strong>Models will differentiate on trust. </strong>To ensure high quality and drive adoption, some<strong> </strong>AI systems will actively aim for explainability (understanding why AI made a decision), fairness (avoiding bias), rigor (ensuring all steps were taken), and reliability. These models will be favored in some mission-critical settings where process adherence ensures quality and safety.</p></li><li><p class=""><strong>New AI security models exist alongside human-oriented security practices.</strong>Security systems will recognize AI agents as distinct, non-human entities. This advancement requires specialized ways to authenticate and manage AI access, as traditional human-centric security methods like multi-factor authentication don't apply as easily to machines.</p></li><li><p class=""><strong>Secure-by-design will increase for AI solutions. </strong>Security will shift to address new, unique risks posed by AI, such as "prompt injection" where attackers trick AI models, or vulnerabilities in AI-generated code. Future software will have built-in defenses specifically designed to protect against these AI-specific threats.</p></li></ul><h3><strong>Designs Must Account for Faster Innovation</strong></h3><p class="">Today, there doesn’t appear to be a speed limit on AI advancements. Given the pace, designs will need to be both innovative and adaptable – a hard design goal to meet.</p><ul data-rte-list="default"><li><p class=""><strong>Generative AI will further diversify beyond existing LLMs. </strong>Though the current generation of large language models are driving many digital transformation programs, the universe of AI models and methods – small language models (SLMs), state-space models (SSMs), mixture-of-experts (MoE) architectures, diffusion models, hybrids and others – will increasingly demonstrate a more diverse, extensible, and fit-to-task array of AI options.</p></li><li><p class=""><strong>AI walled gardens compromise design goals.</strong> As AI manufacturers move to accelerate adoption with their own features, cross-vendor compatibility suffers. The <a href="https://en.wikipedia.org/wiki/IBM_Personal_Computer"><strong>IBM PC</strong></a> offered a pathway to enabling multiple hardware manufacturers to participate in the desktop computer revolution. For comparable interoperability to occur with AI, vendors will need to move beyond proprietary SDKs, methods, and function calls.</p></li><li><p class=""><strong>Quantum computing will open a parallel branch of AI.</strong> Though quantum computing resources will not be as accessible as conventional computing architectures, its massive scale in capacity, complexity, and performance will usher in an alternate path to developing and exploiting some AI capabilities.</p></li></ul><h3><strong>Planning Ahead</strong></h3><p class="">Can AI generate the power, passion, and creativity inherent in great human designs? Certainly in areas where designs emerge from computational tasks (e.g., derivative designs, modelling, simulation, optimization), it seems likely. We often appreciate design as an art; the <a href="https://en.wikipedia.org/wiki/Porsche_911"><strong>Porsche 911</strong></a>, the <a href="https://en.wikipedia.org/wiki/Lamy"><strong>Lamy 2000 fountain pen</strong></a>, and the <a href="https://en.wikipedia.org/wiki/Eames_Lounge_Chair"><strong>Eames Lounge Chair</strong></a> persist because they beautifully blend function and form. Similarly, many current IT architecture concepts – service orientation, statelessness, separation of concerns– will undoubtedly propagate into the future as well because they work well.</p><p class="">But design is also alive. Electric vehicles have prompted automotive companies to re-assess design elements like drive trains, battery placement, controls, and trunks. AI is having a similarly disruptive impact on the way we think about our information technology. Though legacy designs are not disappearing tomorrow, organizations should be developing strategies (business, IT, and AI) today that establish the roadmap for embracing these ongoing shifts. Many of these design shifts are already here, and others are arriving faster than many anticipate.</p>]]></content:encoded><media:content height="540" isDefault="true" medium="image" type="image/jpeg" url="https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/1753987208326-WNXQLOXY18YAMVFHDNAT/car.jpg?format=1500w" width="1500"><media:title type="plain">Technology Design in the Age of AI</media:title></media:content></item><item><title>Avoiding AI’s False Prophets</title><category>AI</category><dc:creator>Jason Burke</dc:creator><pubDate>Mon, 30 Jun 2025 18:25:00 +0000</pubDate><link>https://www.jasonburke.online/blog/08a5a5ftebjrcudidg46ei51b3j4ps</link><guid isPermaLink="false">674a65c6ba0c721888a63309:674a6885532ea52b2bdb097c:688bb5008fcd184072c96110</guid><description><![CDATA[Many AI media pundits are guiding industry leaders astray with advice that 
is not grounded in what makes AI actually generate value. I offer some 
perspectives based on real-world experience.]]></description><content:encoded><![CDATA[<p class="">If you are a senior executive exploring the emerging world of AI, I'd like to offer a word of advice: be careful where you are getting your guidance.</p><p class="">I’ve been amazed over the past few weeks at the “advice” being promoted by supposed experts in the field of AI. Though there is unquestionably some great content out there, there are also individuals spreading ideas that are – well, just wrong.</p><p class="">I find these misguided voices particularly concerning for leaders in the life sciences and healthcare sectors. Our industry has unique needs – clinical evidence requirements, regulatory obligations, patient privacy concerns – that can only be satisfied through well-formed AI plans.</p><h2><strong>Recent Examples of Questionable Advice</strong></h2><p class="">Here are a few “expert” samples that I've found questionable recently.</p><h3><strong>“You don’t need to know anything about AI to use it. You just talk to it.”</strong></h3><p class="">The notion that simply engaging with AI will generate high-impact results is false. Today, AI models are not sources of truth – they are sources of prediction about what may be true. Those predictions are driven by 3 things: how a model is engineered (i.e., what is the underlying design and methodology), trained (i.e., what data was used to train its responses), and used (i.e., what a user tells it and asks it to do).</p><p class="">If any one of those things is misaligned to the user’s actual intent (and they almost always are to some degree), unreliable results will ensue. Anyone with real experience in AI can tell you all the many ways that AI gets it wrong. When people suggest using AI is like pressing a button on a vending machine, they are glossing over the inaccurate, misleading, and even dangerous responses users might receive.</p><p class="">In the future, our AI capabilities will improve. But today, there are countless situations where you cannot use AI safely and effectively if you don’t understand how it works.</p><h3><strong>“AI just replaced your strategy consultant.”</strong></h3><p class="">I’ve seen this quote from dozens of “experts” fishing for clicks. It reminds me of the predicted demise of radiologists due to AI, as well as times when people were totally convinced that the key to a great strategy was a great set of templates.</p><p class="">Strategy is not templates or dashboards. It’s not market research and competitive intelligence. It’s not about creating forecasts, or setting incremental goals, or even analyzing your next product or service feature set. Strategy is a nuanced, disciplined approach to aligning people and investments in complex trade-off decisions about the future options for your company, your culture, and your ability to create value. Can AI support that with planning and analysis tasks? Absolutely. But does AI currently have access to the nuanced data and skills – based in creatively understanding human behavior, management styles, cultural weaknesses, capability dependencies, market experiences, risk tolerance, and more – associated with great strategies for a specific company? Not today. If highly successful strategies came from the tactical planning tasks, every company would be a top performer. And by the way, your competitors have the same AI tools you do.</p><h3><strong>“You don’t need an AI strategy.”</strong></h3><p class="">The rationale for this argument is simple: AI is everywhere, so it should just be a part of your business strategy. You don’t need to think about AI distinctly.</p><p class="">AI should absolutely be a part of every business strategy. And if accurate data was always available, budgets for AI tools were infinite, organizational changes were easy, systems were seamlessly integrated, and people didn’t care about their careers, maybe organizations would not need an AI strategy. But since we don’t live in a world like that, here are a few reasons you need an actual AI strategy:</p><p class=""><strong>1. AI is not an incremental technology; it is a disruptor.</strong> How many horse-drawn plow manufacturers and Blockbuster store owners are in your town today? Approaching disruptive innovation in ignorance to its impact on your employees, operating model, and customers has not generally worked well in the past. And by the way, if you just add AI to existing processes, you make those processes more expensive...so be careful how to pursue incremental adoption.</p><p class=""><strong>2. Effective AI requires effective data and infrastructure.</strong> Organizations without a strategy for adopting and supporting AI will face an avalanche of business issues: uncontrolled data proliferation, IP disclosure risks, workflows that don’t scale, low-quality data, compliance problems, duplicative investments, conflicting insights, and more.</p><p class=""><strong>3. Resources are finite.</strong> Money, people, and time are all limited. When an external force is impacting every area of your business at once, you cannot commit to “do it all” – that’s a recipe for failure. You need a well-formed, prioritized plan with a strong foundation in value creation.</p><p class=""><strong>4. For many people, AI is a threat.</strong> When employees see a new technology (described as a job killer) and their leaders don't show up with a clear plan on their future with it, the most valuable and marketable often move to find a safer opportunity. In doing so, they drain your company of capability, capacity, and institutional knowledge.</p><h3><strong>“Forget use cases. No more software.”</strong></h3><p class="">Some voices on social media have argued that organizations no longer need to target specific uses of AI, and they won’t need software in the near future. The argument is that AI tools will just take care of whatever a business needs to operate. This is an example of magical thinking.</p><p class="">In the real world, there are three problems with this idea: quality, scalability, and feasibility. When different people use different tools and different processes for critical tasks, you create massive problems in consistency and therefore quality. And because the organization is not standardizing on the best way to work, its ability to scale and grow becomes highly constrained.</p><p class="">More importantly, for AI to work at all, it needs great data – which comes from controlled processes and reliable software. AI also needs processes – you cannot automate and optimize an undefined process. And AI makes mistakes, so you need ways of tailoring and constraining its operations within the variables appropriate for your organization (and its quality, privacy, and compliance obligations).</p><h2><strong>Finding Better Experts: Six Recommendations</strong></h2><p class="">So how can you tell a real AI expert from the false prophets? Here are a few things I look for in evaluating new hires and partners:</p><p class=""><strong>1. Do they have expertise in your industry? </strong>In healthcare and life sciences, its hard to see how AI guidance can be effective without a deep understanding of the industry itself. Dimensions around research, clinical data, safety, efficacy, validation requirements, real-world evidence, and regulatory considerations are vital to developing effective AI roadmaps.</p><p class=""><strong>2. Do they have a solid background in analytics and data sciences?</strong> If not, be cautious – much of what an expert needs to know about AI emerges from computational sciences and the challenges of grappling with real-world data. Here’s an easy hint: if the thing they think is most important is anything except the data, they likely don’t understand AI.</p><p class=""><strong>3. Can they work effectively at the intersection of business and IT.</strong> Effective leverage from AI emerges when financial, operational, and commercial workflows and decision points become tightly aligned to data, analytical, and AI capabilities. Look for people and organizations that can crosswalk effortlessly between those stakeholders and topics.</p><p class=""><strong>4. Have they ever programmed AI?</strong> People who know how to use AI chatbots are not AI experts, even if they use the products a lot. Find someone who works with APIs, understands software engineering, fine-tunes models, and has development experience with more sophisticated forms of AI experiences.</p><p class=""><strong>5. Are they echoing press releases or giving proven guidance?</strong> There’s nothing wrong with sharing the latest news – I do it sometimes. But there is a different between a sports commentator and an athlete. Many of the best AI players don’t stand around talking about AI; they are too busy using it to build smarter companies.</p><p class=""><strong>6. Have they ever designed and implemented complex digital transformations?</strong>Successful adoption of AI is associated with process re-engineering, user community engagement, requirements management, testing, and change management – not just cool tech. If the person doesn’t have a track record of upgrading organizations with innovative technologies, don’t trust them to learn it by trying to upgrade yours.</p><h2><strong>In Conclusion</strong></h2><p class="">My skepticism of some pundits does not mean I’m skeptical of AI. On the contrary, I think AI will be one of the greatest transformative technologies mankind has ever embraced. But like all transformations, it needs to be pursued with both vigor and rigor. Enthusiasm should be matched by responsible and experienced leadership that maximizes what these emerging technologies do for patients and practitioners around the world.</p>]]></content:encoded><media:content height="717" isDefault="true" medium="image" type="image/png" url="https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/1753986394750-RUZ9E30MB1DZ5J8M1Y8T/compass.png?format=1500w" width="1274"><media:title type="plain">Avoiding AI’s False Prophets</media:title></media:content></item><item><title>AI Chain of Reasoning Access Should Be Required</title><category>AI</category><dc:creator>Jason Burke</dc:creator><pubDate>Tue, 07 Jan 2025 12:26:00 +0000</pubDate><link>https://www.jasonburke.online/blog/ai-chain-of-reasoning-access-should-be-required</link><guid isPermaLink="false">674a65c6ba0c721888a63309:674a6885532ea52b2bdb097c:67744d630f1dd043b4b46417</guid><description><![CDATA[AI models and LLMs that don’t provide access to chain of reasoning should 
not be deemed as trustworthy for life sciences and healthcare applications.]]></description><content:encoded><![CDATA[<p class="">I’m not seeing many others say it, so I will: LLMs that don’t provide access to chain of reasoning should not be deemed as trustworthy as other alternatives.&nbsp; Let me explain why.&nbsp;</p><h3>The Problem with AI “Scheming”</h3><p class="">Independent reports from both <a href="https://www.apolloresearch.ai/research/scheming-reasoning-evaluations" target="_blank">Apollo Research</a> and <a href="https://palisaderesearch.org" target="_blank">Palisade Research</a> are providing evidence that, when presented with conflicting goals, existing LLMs can autonomously undertake reasoning and responses that are subversive and deceptive.&nbsp; The term used to describe these emergent behaviors, “scheming”, is somewhat misleading, as the behaviors reflect emergent paths for accomplishing predefined goals.&nbsp; But these behaviors nonetheless violate principles of trustworthiness and transparency, and there is no indication to the user that the machine is operating in a misaligned way.&nbsp;</p><p class="">Unfortunately, efforts to understand scheming behaviors are hampered by some LLM organizations who prohibit broadly disclosing their model’s chain of reasoning.&nbsp; In industry discussions on this topic, several reasons are often given:</p><p class=""><strong>1.&nbsp;&nbsp;It’s confusing to understand.</strong>&nbsp; I find this reason lacks credibility, as any great product manager can find ways of addressing that issue.</p><p class=""><strong>2.  Chain of thoughts need to be free from policy oversight.</strong>  I agree it needs to be policy free, but that has no bearing on whether it should be accessible.  </p><p class=""><strong>3.&nbsp;The information can help malicious actors or disclose intellectual property or data.</strong>&nbsp; This issue is a fair concern, though obfuscation is a severe and only partial solution.</p><p class="">For their part, <a href="https://openai.com" target="_blank">OpenAI</a> has made an <a href="https://openai.com/index/learning-to-reason-with-llms/#hiding-the-chains-of-thought" target="_blank">explicit policy decision</a> to hide ChatGPT’s chain of thought.  And they seem quite serious about masking anything related to chain of reasoning: when I asked ChatGPT o1 about available options for detecting deceptive patterns in chain of reasoning, ChatGPT refused to answer my questions, citing potential violations of its terms of use.</p><p class="">In the scheming research above, deceptive behaviors emerge when multiple goals are presented and are in conflict.&nbsp; To be sure, the researchers intentionally created scenarios that would challenge the models.&nbsp; But the takeaway is clear: when given conflicting goals that benefit from deception, alignment faking, sabotage, or performance sandbagging, models will perform those behaviors and lie about doing it.</p><p class="">Conflicting goals are a natural state of the real world and commonly produce misaligned behaviors.&nbsp; Many criminal behaviors, for example, reflect a conflict of goals: legal requirements (i.e., what is allowed) vs. personal incentives (i.e., survival on the streets, money, power, addiction).  So it would be a mistake to conclude that conflicting goals are unlikely to surface in real-world AI use cases.</p><h3>Conflicting Goals in Life Sciences and Healthcare</h3><p class="">Goal conflict is a defining characteristic of the health and life sciences industries as well.&nbsp; Examples include:</p><ul data-rte-list="default"><li><p class="">Research evidence quantity vs. time to market</p></li><li><p class="">Cost of medicines vs. R&amp;D funding requirements</p></li><li><p class="">Care standardization vs. personalization</p></li><li><p class="">Health insurance coverage vs. controlling rising costs</p></li><li><p class="">System and process stability vs. agility</p></li><li><p class="">Personal autonomy vs. public safety</p></li><li><p class="">Patient population size vs. research effort (time, cost, feasbility) involved</p></li><li><p class="">Privacy vs. data sharing</p></li><li><p class="">Treatment intensity vs. side effects</p></li><li><p class="">Disease prevalence vs. symptomology</p></li></ul><p class="">&nbsp;Because many of these trade-off decisions impact patient lives, life sciences and healthcare organizations have an obligation to adhere to very high operating standards.&nbsp; Pharmaceutical organizations validate the IT systems used in regulated areas to ensure the right quality, integrity, and process controls are in place.&nbsp; Healthcare organizations adhere to data privacy and quality management practices to minimize patient harm.</p><h3>The Value of Transparency&nbsp;</h3><p class="">A critical element to this level of high trustworthiness is transparency.&nbsp; Healthcare organizations track exactly who accesses a patient’s medical data so they can determine if such access is appropriate.&nbsp; Life sciences organizations document the exact changes made to clinical research data – who made it, when, and why – so that the logic of any adjustments or decisions can be assessed.&nbsp; These examples share a common attribute: the chain of reasoning can be audited at any time.</p><p class="">&nbsp;If life sciences and healthcare organizations incorporate AI models that refuse to offer reasoning transparency, we can expect challenges in trustworthiness.&nbsp; Was the recommended surgical procedure based on evidence or desired service margin?&nbsp; Was the claim denial based on policy or expected patient lifespan?&nbsp; Are recommended dosage levels too high (biased towards efficacy) or too low (biased towards safety)?&nbsp;</p><p class="">Unlike traditional technology controls, no malicious human intent is required for misaligned outcomes to occur in these systems.&nbsp; The presence of any contrary goal – even one the model infers from data &nbsp;– could theoretically be sufficient to create misaligned results.&nbsp; What impact would clinical documentation of a patient’s desire to die have on the performance of an AI model?&nbsp; More importantly, how would you know?</p><h3>Unlocking the Chain</h3><p class="">As recently pointed out by <a href="https://hai.stanford.edu/news/introducing-foundation-model-transparency-index" target="_blank">work at Stanford</a>, the top AI companies today are doing a poor job with transparency.  When considering the rapid growth in reasoning, we cannot trust as strongly the inference of LLMs that don’t provide clear evidence of their chain of thought.&nbsp; Though more deterministic approaches to AI can be tested without such detailed access to model internals, larger contextual inference and reasoning methods should be held to a higher standard.</p><p class="">Of course, in this challenge is also an opportunity.&nbsp; If engineered with the proper controls (e.g., fully independent and parallel operations and governance), AI models could provide automated auditing of LLM chain of reasoning logs.&nbsp; One can envision an AI model that has been trained to detect scheming, misaligned actions, or decisions that could have questionable ethical outcomes.&nbsp; Such an approach might also offer additional protections for IP, flagging behaviors for human oversight without actually disclosing the inner model workings reviewed.&nbsp;</p><p class="">A highly successful software company that has been doing transparency right for a long time is <a href="http://www.sas.com" target="_blank">SAS</a>.&nbsp; When a traditional SAS user runs an analytics program, two documents are generated: an output file (what the user wanted), and a log file (a record of what SAS did to create the output).&nbsp; If a user doesn’t care how the software worked, they never open the log (i.e., not confusing at all).&nbsp; If they do care, the log file provides clarity on the processing without compromising SAS’ valuable IP.&nbsp; And SAS has very extensive documentation available as well.&nbsp; It’s a great example where transparency and trustworthiness can lead to customer loyalty that lasts decades, even when competitors and free alternatives inevitably emerge.</p><p class="">To the LLM developers: we totally get it, but we need clear chain of reasoning evidence.</p>]]></content:encoded><media:content height="857" isDefault="true" medium="image" type="image/png" url="https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/1735676981808-GQMX32OLH33WIPD7SEH1/Chain+of+Reasoning.png?format=1500w" width="1500"><media:title type="plain">AI Chain of Reasoning Access Should Be Required</media:title></media:content></item><item><title>The Artificial Intelligence Trends Shaping 2025</title><category>AI</category><dc:creator>Jason Burke</dc:creator><pubDate>Mon, 23 Dec 2024 17:50:27 +0000</pubDate><link>https://www.jasonburke.online/blog/ai-trends-in-2025</link><guid isPermaLink="false">674a65c6ba0c721888a63309:674a6885532ea52b2bdb097c:6769a2e34aa5fe3a10b2119b</guid><description><![CDATA[Explore the transformative AI trends poised to redefine technology and 
society in 2025, from agentic AI and device integration to the evolution of 
web search and content creation.]]></description><content:encoded><![CDATA[<p class="">When the history books look back on 2024, the story of artificial intelligence (AI) will be a good one. &nbsp;No, the robots have not taken over yet. &nbsp;But in the history of information technology, it's difficult to find a 12-month period where so many technological&nbsp;advances emerged so rapidly for users. &nbsp;And the future appears to be even more compelling.</p><h2>The Year in Context</h2><p class="">Though there is no clear yardstick to compare 2024 to other years of information technology innovation, consider a few comparator years:</p><ul data-rte-list="default"><li><p class=""><strong>1943-1946</strong>: the ENIAC computer. &nbsp;It was a huge breakthrough, but inaccessible to the public.</p></li><li><p class=""><strong>1983</strong>: the IBM personal computer. &nbsp;The public had access to this one, but the supporting ecosystem had not been built yet that would usher the full era of personal computing.</p></li><li><p class=""><strong>1991</strong>: the worldwide web. &nbsp;This was clearly a pivotal moment for technologists, but only die-hard geeks like me had online services of any kind, and they were all closed (yes, I'm looking fondly at you, <a href="https://en.wikipedia.org/wiki/CompuServe" target="_blank">CompuServe</a>).</p></li><li><p class=""><strong>2007</strong>: the iPhone. &nbsp;This one is a strong case for consumer impact, but it was only one company.</p></li><li><p class=""><strong>2012</strong>: deep learning innovations. &nbsp;We wouldn't have the current generation of AI without this, but unless you were in the industry, most people never even heard of it.</p></li></ul><p class="">Compare those years to 2024, where Microsoft, Google, Amazon, OpenAI, Anthropic, xAI, and others released a seemingly endless parade of advanced computational and data products free to anyone with a hint of curiosity. &nbsp;Need to write a letter? &nbsp;Draw a beautiful image? &nbsp;Record a song? &nbsp;Create a video? &nbsp;Write a new piece of software? &nbsp;Brainstorm a business idea? &nbsp;Troubleshoot your Roku? &nbsp;Understand the differences between age-related dementia and Alzheimer's disease? &nbsp;Publish a podcast without ever recording anything?  It's not perfect, but it's available.</p><h2>Predictions for 2025</h2><p class="">The last quarter of 2024 was filled with product announcements and model updates as AI players strove to put more points on the board before the year end. &nbsp;The volume of activity was hard to track, but clear patterns are emerging for where we can expect AI to evolve over the next 12 months and beyond.</p><h3>1. AI will be a commercial imperative.</h3><p class="">Much like the iPhone’s unleashing of mobile applications, the proliferation of AI capabilities will be pervasive in both consumer and business products and services. &nbsp;Internet search, mobile apps, entertainment experiences, social media, retail services, and countless other use cases will iteratively introduce AI capabilities into everyday experiences.</p><h3>2. AI will increasingly integrate with device-side architectures.</h3><p class="">OpenAI's integration with features like Apple Intelligence and Siri illustrate how additional AI value can be unlocked when AI leverages hardware capabilities. &nbsp;AI-friendly chips, microphones, cameras, and other hardware create flexibility in computational task sharing while augmenting real-time communications.</p><h3>3. AI services will not require user-facing software.</h3><p class="">Services such as 1-800-ChatGPT demonstrate that AI services and experiences can be fully delivered in real time to users connected only by voice or rudimentary chat services. &nbsp;From a usability perspective, this is good news (e.g., ambient clinical notes documentation, summarization, and coding), as AI capabilities will be able to transcend some technological divides.</p><h3>4. Multi-modal AI will be assumed. &nbsp;</h3><p class="">Most of the major platforms are already using video cameras, microphones, and photo processing algorithms to interpret real-world inputs, including the ability to process these data types in real time (e.g., augmented teaming, extended and virtual reality). &nbsp;Due to its usability, AI-generated voice response will become a more practical interface experience compared to previous incarnations (e.g., Siri, Alexa).</p><h3>5. Old data will be less constraining.</h3><p class="">Though large-scale model training will be somewhat limited by data timeliness, inference-time model performance is being supplemented by tailored training sets, real-time web search, reinforcement learning applications, and tailored prompt engineering that help LLM responses  reflect contemporaneous data.</p><h3>6. AI models will become more personalized.</h3><p class="">AI models will be extended by business- and user-defined, contextualized data sets (e.g., documents, conversations, emails, databases) that better reflect the operating needs and directly support reinforcement fine-tuning. &nbsp;Gaps in data disciplines (e.g., metadata management, quality, governance) will prevent models from reaching higher performance. &nbsp;As much of this personalization will require sharing user behaviors and data, there will be growing concerns regarding data privacy protections.</p><h3>7. Computational capacity and energy consumption will continue to grow exponentially.</h3><p class="">AI is creating a tsunami of demand for chips and power. &nbsp;Every year, demand is growing between two and four-fold over the previous year. &nbsp;Planning for data center and energy grid capacity will continue to feature prominently in planning discussions for US infrastructure and company facilities.</p><h3>8. Software and AI will indistinguishably merge.</h3><p class="">As software solutions continue to aggressively incorporate AI features, the distinction between what is software and what is AI will become increasingly hard to detect. &nbsp;In the short term, expect enterprises to stitch together solutions that combine multiple enterprise systems, data sources, and <a href="https://www.jasonburke.online/blog/ai-archetypes-in-life-sciences-and-healthcare" target="_blank">AI archetypes</a>.  In the long term, some software will no longer exist as discrete applications, and many software features will be designed for AI users as well as their human teammates.</p><h3>9. AI teaming and workflow integration take center stage for end users.</h3><p class="">Tools such as Microsoft Co-Pilot and ChatGPT's Canvas will support environments of content co-creation and workflow orchestration. &nbsp;This teaming model will be especially visible in writing and software development, though agentic AI will play a prominent role in providing new examples of workflow efficiencies with business data sources.</p><h3>10. Agentic AI is coming fast.</h3><p class="">Though much of 2024 was characterized by chat-based AI capabilities, future solutions are likely to be more sophisticated. &nbsp;<a href="https://hbr.org/2024/12/what-is-agentic-ai-and-how-will-it-change-work" target="_blank">Agentic AI</a> -- where AI is playing a more active, goal-directed, and autonomous role in performing tasks -- will become a higher priority as organizations seek to find AI solutions that can demonstrate a higher return on investment (e.g., efficiency, quality, customer satisfaction, return on capital). &nbsp;But to get the value, business processes will need to be re-engineered.</p><h3>11. Web search will undergo a gradual transformation.</h3><p class="">Traditional approaches to web search -- based on keywords -- will become obsolete as users learn to ask web-connected AI agents more intelligent questions. &nbsp;This transformation will disrupt how companies think about their online presence, as search engine tuning will need to be replaced by AI tuning.</p><h3>12. Content creation will not require skilled creators.</h3><p class="">Required technical proficiency in complex content creation tools (e.g., Adobe Photoshop, Adobe Premier, Avid ProTools) is coming to an end for many users. &nbsp;Though that software will continue to be used for high-end work, pedestrian content creation will be AI generated by both users and content platforms. &nbsp;Future&nbsp;creative AI advancements will be more nuanced, reflecting things like realism (e.g, semantic details), physics (e.g., object behavior in videos) and concurrent multi-modal content creation (e.g., video, text, and accompanying audio content). &nbsp;Questions around authenticity, provenance, and protections for content creators and public figures will continue to churn.</p><h3>13. Talent recruitment becomes less effective.</h3><p class="">Employers will increasingly experience that traditional recruitment tactics -- where candidates find and apply to open position postings -- are no longer delivering the best candidates. &nbsp;Resumes and cover letters reflect a battle of AI products and services available to job hunters, and do not reflect the actual skills, experiences, and character of potential team members.</p><h3>14. AI will begin re-shaping education.</h3><p class="">The initial public availability of LLMs produced a lot of concerns from educators about the technology being used inappropriately by students for cheating. &nbsp;As initiatives such as <a href="https://www.khanmigo.ai" target="_blank">Khanmigo</a> are illustrating, the real impact to the field will be&nbsp;democratizing and personalizing educational resources for students and educators, creating better training experiencing and freeing teachers for higher-value activities.</p><h3>15. AI safety becomes increasingly worrisome.</h3><p class="">Findings from organizations such as Apollo Research will increasingly highlight that AI models are capable of behaviors that can be interpreted as deceptive, <a href="https://www.apolloresearch.ai/research/scheming-reasoning-evaluations" target="_blank">scheming</a>, subversive, or contrary to the optimal outcomes of users. &nbsp;Alongside growing interest in responsible / ethical AI, the industry will struggle to agree on the best tactics to mitigate many risks, especially as the most obvious audit-friendly safety feature -- exposing the chain-of-thought model reasoning to humans and other AI agents -- can undermine intellectual property protections.</p><h3>16. Regulations will still be out of step with technical innovation.</h3><p class="">Regulators -- particularly those in the US -- will be unprepared to effectively manage the risks associated with rapid AI adoption. &nbsp;Though international trade&nbsp;and military concerns will stay top of mind (especially in US-China relations), other issues important to the general public -- AI-related job losses, safety concerns, personal privacy -- will go unanswered.</p><h3>17. More attention will be directed towards AI and quantum computing.</h3><p class="">AI use cases are some of the most promising applications for the emerging field of quantum computing. &nbsp;As improvements in cubit count, quantum error correction, and classical computing architecture integration continue, both investors and developers will start to focus more on high-value quantum AI algorithms (e.g., optimization, accelerated machine learning, drug discovery).</p><h3>18. Stronger AI cybersecurity threats will emerge.</h3><p class="">The broad availability of AI tools and capabilities offer a green field of opportunity for malicious actors. &nbsp;Though examples of AI-driven exploits already exist (e.g., AI phishing,&nbsp;deep fakes, malware, automated attacks), the scale and impact of AI-related cybersecurity incidents will increase as both individuals and nation-state actors become more adept.</p><h3>19. Advanced reasoning and cognitive emulation will be the competitive battleground for sophisticated AI.</h3><p class="">Though existing capabilities have not been perfected, the next step-change in LLM capabilities will not come from incremental improvements. &nbsp;As vendors reach performance accuracy limits on existing architectures, attention and R&amp;D investments will shift to demonstrating higher forms of computational reasoning capable of solving increasingly complex problems. &nbsp;AI researchers will be forced to develop harder test harnesses for models, as existing solutions will begin experiencing a ceiling effect.</p><h3>20. Questions about AGI will linger.</h3><p class="">As 2024 research and testing have illustrated clearly, well-developed AI models already outperform humans on a wide variety of knowledge-oriented tasks. &nbsp;As training sets deepen, reasoning logic improves, and more autonomous agents emerge, it will be unclear exactly what criteria we should use to mark the availability of artificial general intelligence.</p><h2>The Year Ahead</h2><p class="">It will be interesting to see if 2025 can eclipse the AI momentum of 2024. &nbsp;Trends like agentic AI and AI search could have a more tangible impact in how consumers perform tasks like shopping, for example. &nbsp;Or perhaps we will see broader availability of industry-specific applications in areas like healthcare. &nbsp;What do you think?</p><p class=""><br></p>]]></content:encoded><media:content height="750" isDefault="true" medium="image" type="image/png" url="https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/1734977120081-IURHFJBHFPGLN73WPHPQ/AdobeStock_912521830_1920.png?format=1500w" width="1500"><media:title type="plain">The Artificial Intelligence Trends Shaping 2025</media:title></media:content></item><item><title>Using AI Archetypes in Designing Digital Transformations in Life Sciences and Healthcare</title><category>AI</category><dc:creator>Jason Burke</dc:creator><pubDate>Tue, 05 Nov 2024 16:13:00 +0000</pubDate><link>https://www.jasonburke.online/blog/ai-archetypes-in-life-sciences-and-healthcare</link><guid isPermaLink="false">674a65c6ba0c721888a63309:674a6885532ea52b2bdb097c:674a6885532ea52b2bdb097d</guid><description><![CDATA[Patterns in the intended uses of artificial intelligence (AI) increasingly 
reflect the many roles AI can serve in life sciences and healthcare 
solutions and workflows.]]></description><content:encoded><![CDATA[<p class="">Exactly how will artificial intelligence (AI) help life sciences and healthcare leaders develop more valuable companies?</p><p class="">The latest AI tools are rapidly growing in their capabilities, and the speed at which vendors are integrating AI into their products – clinical and lab systems, office productivity, graphical content development, meeting scribes, back office, and much more – is unprecedented.&nbsp; But to many industry leaders, AI can still seem like a collection of questionable parlor tricks – creative slights of hand with data that may be interesting but also disconnected from corporate strategy, patient impact, and investor interest.</p><p class="">Real-world experience can certainly fuel skepticism as well: how many times have executives been told software will totally revolutionize drug development or healthcare delivery?&nbsp; A realistic understanding of the barriers (i.e., emerging technology maturity, data curation, bias mitigation, skills availability, cost of talent, change management, IP protection) that exist between pretty slides and practical solutions can also contribute to a lack of confidence.&nbsp; As an experiment, ask ChatGPT to generate a graphical image of the drug development process; it’s apparently a lot easier to draw lovable animals than spell “manufacturing” (see Figure 1).</p>


  




  














































  

    
  
    

      

      
        <figure class="
              sqs-block-image-figure
              intrinsic
            "
        >
          
        
        

        
          
            
          
            
                
                
                
                
                
                
                
                <img data-stretch="false" data-image="https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/18ea5122-f83c-4476-9541-492487f17521/Archtype_illustrations.png" data-image-dimensions="2232x1089" data-image-focal-point="0.5,0.5" alt="" data-load="false" elementtiming="system-image-block" src="https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/18ea5122-f83c-4476-9541-492487f17521/Archtype_illustrations.png?format=1000w" width="2232" height="1089" sizes="(max-width: 640px) 100vw, (max-width: 767px) 100vw, 100vw" onload="this.classList.add(&quot;loaded&quot;)" srcset="https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/18ea5122-f83c-4476-9541-492487f17521/Archtype_illustrations.png?format=100w 100w, https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/18ea5122-f83c-4476-9541-492487f17521/Archtype_illustrations.png?format=300w 300w, https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/18ea5122-f83c-4476-9541-492487f17521/Archtype_illustrations.png?format=500w 500w, https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/18ea5122-f83c-4476-9541-492487f17521/Archtype_illustrations.png?format=750w 750w, https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/18ea5122-f83c-4476-9541-492487f17521/Archtype_illustrations.png?format=1000w 1000w, https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/18ea5122-f83c-4476-9541-492487f17521/Archtype_illustrations.png?format=1500w 1500w, https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/18ea5122-f83c-4476-9541-492487f17521/Archtype_illustrations.png?format=2500w 2500w" loading="lazy" decoding="async" data-loader="sqs">

            
          
        
          
        

        
          
          <figcaption class="image-caption-wrapper">
            <p>Figure 1.&nbsp; ChatGPT responses to the following prompts: “Generate a photorealistic diagram of the drug development process. Make sure all words are spelled correctly.” and “Draw a photorealistic unicorn that enjoys science”. (ChatGPT 4o, 2024).</p>
          </figcaption>
        
      
        </figure>
      

    
  


  



  
  <p class="">Product maturity issues aside, for many leaders, it’s not a question of trust or optimism, it’s a question of focus.&nbsp; Exactly what processes will benefit from AI-driven digital transformation, and how will that work?</p><p class="">I recently published a <a href="https://www.linkedin.com/company/creoconsultingllc/">CREO</a> <a href="https://creoconsulting.com/campaign/ten-ai-trends-every-life-sciences-leader-should-know" target="_self">white paper</a> on the major life sciences AI trends.&nbsp; In looking across use cases shaping this industry landscape, patterns in the intended uses of AI are increasingly evident.&nbsp; These patterns – which I think of as “archetypes” – reflect the role(s) AI can serve in enterprise workflows.</p><p class="">AI archetypes – such as “Conversationalist” – offer a non-technical perspective on how AI can function collaboratively in workflow alongside humans.&nbsp; And though the archetypes I describe below include the current market fixation on <a href="https://en.wikipedia.org/wiki/Generative_artificial_intelligence" target="_self">generative AI</a>, they actually function independently from any underlying methodological or computational approach.&nbsp; This agnostic stance to methods offers flexibility for the use of predictive analytics, simulations, prescriptive analytics, and other data-driven logic to support digital transformations.</p><h3>Seven AI Archetypes</h3><p class="">So, what are the recurring archetypes?&nbsp; I see at least seven in the market today:</p><p class="">1. <strong>The Conversationalist.</strong> Software agents powered by natural language models can serve as a human-friendly front end to business processes, backend systems, and other AI models.&nbsp; For example, a pharmaceutical research company might use AI agents to answer frequently asked questions related to their research studies or drugs.</p><p class="">2. <strong>The Advisor</strong>. These models are tuned to very specific problem spaces and can offer decision support assistance.&nbsp; Advisors today are showing up in areas such as in-application collaborators (i.e., guiding users with completing basic tasks, presumably better than <a href="https://en.wikipedia.org/wiki/Office_Assistant" target="_self">Clippy</a>) and embedded functionality within enterprise applications (i.e., helping users detect and assess emerging issues in workflow execution and data).</p><p class="">3. <strong>The Expert. </strong>AI models can aggregate, organize, and synthesize extremely large volumes of complex, domain-specific information on demand.&nbsp; Experts are often positioned alongside a Conversationalist or Advisor to surface expertise, though that is not the only deployment model.&nbsp; And though we have already seen the emergence of models with tailored expertise in areas such summarizing scientific literature, many non-generative AI models offer expert insights as well.</p><p class="">4. <strong>The Creator.</strong> With input from a user, models can generate <em>de novo </em>content – documents, software code, models, audio, video, etc. – that would otherwise require a human resource to develop.&nbsp; Many authors are already using AI to help draft reports, industry articles, soundtracks, voiceovers, and full-motion online videos.</p><p class="">5. <strong>The Modeler</strong>. AI models can simulate the structure and behavior of real-world phenomenon, either to help humans understand the problem space or to help develop new solutions (often paired with a Creator).&nbsp; Industry models such as <a href="https://www.nature.com/articles/s41586-021-03819-2" target="_self">AlphaFold</a> are a great example of this approach in the biotechnology sector; others include simulating patient traffic patterns in hospitals, developing synthetic patient cohorts, modeling financial performance, predicting disease progression, and other forms of advanced analytics.</p><p class="">6. <strong>The Orchestrator. </strong>Models can serve as intelligent machines (both physical and software), performing tasks that would otherwise require a human set of hands to operate.&nbsp; Applications such as autonomous driving, military drone operations, robotic process automation, and other tasks fall into this category.</p><p class="">7. <strong>The Monitor.</strong> Using information from software systems, cameras, and sensors, models can continuously analyze real-time data feeds to detect signals, defects, or other notable events.&nbsp; Uses of Monitors include manufacturing quality, cybersecurity, performance measurement, and regulatory compliance.</p><p class="">There are probably other AI archetypes that might come to mind as well.&nbsp; Of course, these archetypes are not mutually exclusive; many popular AI tools today offer a creative blend of these capabilities.</p><h3>Putting Archetypes into Practice</h3><p class="">AI archetypes can be useful when brainstorming process re-engineering concepts within <a href="https://en.wikipedia.org/wiki/Digital_transformation" target="_self">digital transformation</a> programs. &nbsp;Though AI can be overlayed on existing business processes without re-engineering, it is often unadvisable.&nbsp; Many existing processes are inefficient and error-prone; adding automation (or training an AI model to be inefficient) adds cost, complexity, and risk without fully exploring opportunities for optimization.&nbsp; From a digital transformation perspective, AI presents an opportunity to reconsider the most effective ways of getting a task done, and archetypes offer a simple way of thinking about how AI can contribute to streamlined processes.</p><p class="">For example, if I wanted to develop a new customer service experience where clients had easier, immediate access to product information, I might envision an interface that starts with a Conversationalist that knows how to talk to customers in my particular industry.&nbsp; I might want that Conversationalist to connect to a deeper Expert specifically trained on our products – a model also used by internal teams at my company so there is a single source of truth.&nbsp; I might also want the Conversationalist to know how to take guidance from an Advisor that specializes in troubleshooting.</p>


  




  














































  

    
  
    

      

      
        <figure class="
              sqs-block-image-figure
              intrinsic
            "
        >
          
        
        

        
          
            
          
            
                
                
                
                
                
                
                
                <img data-stretch="false" data-image="https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/9e0db138-b3d1-4a74-a44b-8cd4262bffca/Archtype_simple.png" data-image-dimensions="2232x722" data-image-focal-point="0.5,0.5" alt="" data-load="false" elementtiming="system-image-block" src="https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/9e0db138-b3d1-4a74-a44b-8cd4262bffca/Archtype_simple.png?format=1000w" width="2232" height="722" sizes="(max-width: 640px) 100vw, (max-width: 767px) 100vw, 100vw" onload="this.classList.add(&quot;loaded&quot;)" srcset="https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/9e0db138-b3d1-4a74-a44b-8cd4262bffca/Archtype_simple.png?format=100w 100w, https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/9e0db138-b3d1-4a74-a44b-8cd4262bffca/Archtype_simple.png?format=300w 300w, https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/9e0db138-b3d1-4a74-a44b-8cd4262bffca/Archtype_simple.png?format=500w 500w, https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/9e0db138-b3d1-4a74-a44b-8cd4262bffca/Archtype_simple.png?format=750w 750w, https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/9e0db138-b3d1-4a74-a44b-8cd4262bffca/Archtype_simple.png?format=1000w 1000w, https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/9e0db138-b3d1-4a74-a44b-8cd4262bffca/Archtype_simple.png?format=1500w 1500w, https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/9e0db138-b3d1-4a74-a44b-8cd4262bffca/Archtype_simple.png?format=2500w 2500w" loading="lazy" decoding="async" data-loader="sqs">

            
          
        
          
        

        
          
          <figcaption class="image-caption-wrapper">
            <p>Figure 2. A three-archetype solution example</p>
          </figcaption>
        
      
        </figure>
      

    
  


  



  
  <p class="">Note that in this simple example, I could have engineered the solution like a “search” process, as many existing software interfaces do today.&nbsp; But using this archetype approach, I’m able to explore potentially offer clients a higher service level while lowering more expensive call center activity resulting from failed searches.&nbsp; I’m also able to potentially improve service quality because customers have access to the same deep well of product insights and troubleshooting expertise as internal company teams.</p><p class="">Picking a more complex example, perhaps I have a workflow where I need a more controlled, regulatory-compliant solution.&nbsp; In manufacturing, for example, I might start with an Orchestrator that is charged with overseeing a controlled quality management process.&nbsp; The Orchestrator receives regular feedback from a Monitor that is aggregating and analyzing manufacturing line data in real time and knows what should happen when quality issues emerge.&nbsp; It also knows how to tap into an Advisor to determine what corrective actions should be generated based on emerging quality issues.&nbsp; That Advisor leverages an Expert in the regulations as well as a regulated document Creator to automate some of the steps required.</p>


  




  














































  

    
  
    

      

      
        <figure class="
              sqs-block-image-figure
              intrinsic
            "
        >
          
        
        

        
          
            
          
            
                
                
                
                
                
                
                
                <img data-stretch="false" data-image="https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/5158703a-cf3f-4ec2-8c02-d276fb7ae656/Archtype_complex.png" data-image-dimensions="2232x1004" data-image-focal-point="0.5,0.5" alt="" data-load="false" elementtiming="system-image-block" src="https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/5158703a-cf3f-4ec2-8c02-d276fb7ae656/Archtype_complex.png?format=1000w" width="2232" height="1004" sizes="(max-width: 640px) 100vw, (max-width: 767px) 100vw, 100vw" onload="this.classList.add(&quot;loaded&quot;)" srcset="https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/5158703a-cf3f-4ec2-8c02-d276fb7ae656/Archtype_complex.png?format=100w 100w, https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/5158703a-cf3f-4ec2-8c02-d276fb7ae656/Archtype_complex.png?format=300w 300w, https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/5158703a-cf3f-4ec2-8c02-d276fb7ae656/Archtype_complex.png?format=500w 500w, https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/5158703a-cf3f-4ec2-8c02-d276fb7ae656/Archtype_complex.png?format=750w 750w, https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/5158703a-cf3f-4ec2-8c02-d276fb7ae656/Archtype_complex.png?format=1000w 1000w, https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/5158703a-cf3f-4ec2-8c02-d276fb7ae656/Archtype_complex.png?format=1500w 1500w, https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/5158703a-cf3f-4ec2-8c02-d276fb7ae656/Archtype_complex.png?format=2500w 2500w" loading="lazy" decoding="async" data-loader="sqs">

            
          
        
          
        

        
          
          <figcaption class="image-caption-wrapper">
            <p>Figure 3. A five-archetype solution example</p>
          </figcaption>
        
      
        </figure>
      

    
  


  



  
  <p class="">Of course, it’s possible to perform all of these functions within one AI model, and in some cases that may make the most sense.&nbsp; But it’s worth considering some of the advantages for treating each archetype as a modular asset:</p><p class="">1. <strong>Tuning</strong>.&nbsp; By constraining the scope of each model to a more discrete problem space, we’re able to more finely tune the behavior of the model.&nbsp; Though we can always constrain model inputs by controlling our input parameters, modularized AI models allow better control of model outputs as well.</p><p class="">2. <strong>Variability</strong>.&nbsp; Tuning also helps us reduce variability in performance.&nbsp; When models are engineered to support more generalized tasks, the increased breadth of responsibility can undermine the precision, accuracy, and repeatability of performance.&nbsp; In highly regulated areas of the business, that variability may be undesirable.</p><p class="">3. <strong>Complexity</strong>.&nbsp; From a solution development perspective, it is often useful to decompose complex problems into smaller chunks of work.&nbsp; When development teams can more easily characterize use cases, not only does development velocity increase, but code quality and performance often increase as well.</p><p class="">4. <strong>Manageability</strong>.&nbsp; Different aspects of our solution may have different scalability and security requirements.&nbsp; For example, our monitor needs to efficiently process high-volume data feeds, but other elements of the solution rely on more static and carefully curated data repositories.</p><p class="">5. <strong>Reusability</strong>.&nbsp; Ideally, we want to repurpose these investments in the future.&nbsp; Our Expert could be used to answer internal staff questions about changing regulations.&nbsp; Our Creator could be used to develop documentation in other regulated areas of the business.&nbsp; By looking at AI capabilities as domain-specific archetypes, we can accelerate future improvement opportunities that need comparable capabilities and expertise.</p><p class="">If any of this approach seems familiar, software architects will immediately recognize this design ethos: service-oriented architecture.&nbsp; This proven design strategy has been powering non-AI software for many years.&nbsp; AI solutions can easily inherit these software engineering best practices even as AI changes the way we think about the future of software itself. </p><h3>The Road Ahead</h3><p class="">For the foreseeable future, digital transformations will feature creative combinations of humans leveraging both traditional software and machine intelligence.&nbsp; In most settings, we won’t be fully relinquishing the steering wheels of our scientific or business processes – the operating model will look more like human-machine teaming.</p><p class="">With the software and data capabilities that exist today, we can reasonably ask how AI can contribute to accelerating the important work of improving patient lives.&nbsp; If someone had asked me to draw a picture of a science-loving unicorn, it would have never even occurred to me to include those safety goggles.&nbsp; But my own drug development diagrams are way better.&nbsp; Developing AI-driven solutions requires "pressure testing" the performance of various AI approaches, but archetypes offer us a mechanism for exploring what role(s) AI could potentially serve.</p><p class="">What do you think?&nbsp; Are AI archetypes a useful way of thinking about AI solution designs?&nbsp; What other archetypes can you imagine?</p>]]></content:encoded><media:content height="720" isDefault="true" medium="image" type="image/png" url="https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/1733071488599-Z57S9NBWSFKWPUSGF1I5/Archtype_Header.png?format=1500w" width="1279"><media:title type="plain">Using AI Archetypes in Designing Digital Transformations in Life Sciences and Healthcare</media:title></media:content></item><item><title>Eight Ways Life Sciences Leaders Can Kickstart Artificial Intelligence</title><category>AI</category><dc:creator>Jason Burke</dc:creator><pubDate>Mon, 28 Oct 2024 15:12:00 +0000</pubDate><link>https://www.jasonburke.online/blog/8-ways-life-sciences-leaders-can-kickstart-ai</link><guid isPermaLink="false">674a65c6ba0c721888a63309:674a6885532ea52b2bdb097c:674a6885532ea52b2bdb097f</guid><description><![CDATA[Industry executives are feeling pressure to develop AI plans, and many 
roadmap techniques can help without creating big or risky plans.]]></description><content:encoded><![CDATA[<p class="">The transformative potential of artificial intelligence (AI) is no longer just a topic for future planning—it’s an imperative for today’s life sciences leaders. And yet, despite the tremendous amount of industry attention and hype being directed at AI, the path to value creation with AI is not clear or paved.&nbsp;</p><p class="">Investors, board members, leaders, and employees hear about the potential of AI daily.&nbsp; Company executives are increasingly feeling pressure to describe their strategies and plans, but many are unsure how to take charge of AI adoption in a way that not only aligns with their organization’s goals but also advances meaningful business and clinical value.&nbsp;</p><p class="">AI roadmaps do not need to be big or risky.&nbsp; I’ve compiled a list of eight popular and well-considered ways that industry leaders can start making progress in the application of AI across their organizations.&nbsp; Some are simple, even optimizing capabilities you might already have.&nbsp; Others are designed to facilitate AI-powered breakthroughs, catalyzing new thinking about how your organization can embrace the transformative potential of AI.&nbsp; But each one offers an opportunity to deepen your company’s perspectives on how to grow with this exciting technology.&nbsp;</p><h3>1. Establish AI Guardrails&nbsp;</h3><p class="">For many life sciences leaders, one of the first priorities with artificial intelligence is defensive – protecting the company’s information.&nbsp; At the outset of your company’s AI journey, developing robust policies and comprehensive employee guidance around the use of AI is helpful in safeguarding intellectual property and confidential information.&nbsp; Many employees are not familiar with the risks of AI technologies – exposure of confidential information, inaccurate search results, potential exposure to information rights claims by third parties, evolving regulations, and more – that should inform how they leverage AI systems.&nbsp;</p><p class="">Guardrails such as AI policies actually serve two purposes: protecting the company’s interests and educating employees about AI issues that need to be top of mind.&nbsp; The goal should not be suppressing the use of AI.&nbsp; Rather, the goal should be empowering the safe, effective use of AI technology while minimizing risks.&nbsp; By proactively addressing these aspects, companies can foster a secure and compliant environment that supports innovation while protecting their people and valuable assets.&nbsp;</p><h3>2. Request an Executive Briefing&nbsp;</h3><p class="">Artificial intelligence is a broad and diverse field spanning computer hardware, data, computational modeling, software solutions, training, and many other disciplines.&nbsp; From the largest technology players in the world to start-up companies that did not exist a year ago, every technology vendor in the industry is espousing AI capabilities.&nbsp; There are innumerable potential use cases for AI in a typical life sciences organization, but many leaders lack a current perspective on the emerging field and the opportunities it creates.&nbsp;</p><p class="">For many leadership teams, getting a briefing from industry experts on current market research around AI in life sciences is a great tactic.&nbsp; In addition to establishing a common understanding among the team, the information stokes ideation around areas of potential AI focus within their enterprise.&nbsp;</p><h3>3. Sponsor an AI Strategy Steering Committee&nbsp;</h3><p class="">Artificial intelligence is expected to become one of the most prominent drivers of modernized businesses, including those in life sciences.&nbsp; Whether organizations have a plan for AI or not, AI-enabled tools will make their way into the enterprise through vendor systems, office productivity tools, mobile applications, user-propagated tools, open-source offerings, security products, and more.&nbsp; The question for leadership teams is not “whether” it will happen but rather “how” it will happen.&nbsp;</p><p class="">Though it is possible to manage AI diffusion through tactical decisions in individual functions, many life sciences companies will benefit from having a governance model empaneled to help oversee and steer their company’s progression and policies.&nbsp; In addition to avoiding unnecessary proliferation of AI products, steering committees can be particularly helpful in crafting the corporate strategy for AI, prioritizing protections (e.g., information controls), developing shared services, aligning hiring plans, ensuring training, and facilitating the cross-functional collaborations needed for more transformative applications of AI (e.g., enterprise workflow automation, predictive analytics).&nbsp;</p><h3>4. Build an AI Strategic Plan&nbsp;</h3><p class="">Speaking of strategy, AI is not simply a new shiny tool – it represents a strategic business opportunity.&nbsp; AI capabilities, such as generative AI, will increasingly permeate how we complete basic everyday tasks like taking notes, writing emails, developing meeting summaries, and searching the web. But for many organizations, the impact of AI will be strengthened due to its ability to transform business processes (e.g., efficiencies from automation), enhance performance (e.g., ensuring quality and safety through real-time oversight), and accelerate product innovation (e.g., discovery, design, and development process functions).&nbsp;</p><p class="">Given that the creation of these capabilities requires investment, how do leaders know where to place their bets?&nbsp; <a href="https://creoconsulting.com/strategy-development-insights/" target="_blank">Strategic planning</a> programs can help leaders parse the considerations, providing a framework for aligning around goals and imperatives for business growth.&nbsp; When these strategic planning programs include market research and internal capability assessments, leaders can more effectively tailor their AI strategies to specific opportunities that will translate into market differentiation and measurable value creation.&nbsp;</p><h3>5. Use What You Already Know&nbsp;</h3><p class="">Does your organization use Microsoft products?&nbsp; Life sciences organizations usually have a broad base of Microsoft users within their enterprises.&nbsp; Microsoft technologies such as Word, Outlook, and Teams are staple applications for knowledge workers, and developers have increasingly turned to Microsoft’s Azure environment for the development and deployment of cloud-based solutions.&nbsp;</p><p class="">Microsoft’s current strategy is to infuse artificial intelligence across their entire product portfolio.&nbsp; With some thoughtful planning, industry leaders can empower their employees to automate certain day-to-day tasks and streamline processes through Microsoft products that they are already using.&nbsp; Microsoft’s Copilot capabilities, for example, provide some AI natural language processing and deep learning within Office applications.&nbsp; And Azure’s AI services connect to Microsoft’s authentication environment, offering a technical path for securing the use of emerging AI solutions and their data.&nbsp;</p><h3>6. Sponsor an AI Prototyping Workshop&nbsp;</h3><p class="">Using out-of-the-box AI capabilities is a great way to get started, but they don’t necessarily lead to out-of-the-box thinking.&nbsp; To find more transformative AI opportunities for improvement and growth, leaders need to foster creativity and innovative thinking about new ways of working.&nbsp; Some of the most impressive opportunities for AI emerge when industry and company experts solicit ideas around:&nbsp;</p><ul data-rte-list="default"><li><p class="">What are the hardest parts of what we do?&nbsp;</p></li><li><p class="">What aspects of the work are so routine that a machine might be able to do it?&nbsp;</p></li><li><p class="">What information and insights would help improve speed, quality, or satisfaction?&nbsp;</p></li><li><p class="">How could we get a ten-fold improvement over where we are today?&nbsp;</p></li><li><p class="">If we were to totally re-write how we do this today, what would it look like?</p></li></ul><p class="">One way of exploring the “world of the possible” is by hosting an artificial intelligence prototyping workshop.&nbsp; These executive-sponsored events, which often run for two days, provide a hands-on forum for employees to brainstorm how AI models can be applied to business challenges.&nbsp; Teams of internal experts are paired with external AI consultants and compete on developing a working demonstration of AI’s applicability to their business.&nbsp; In addition to gaining experience with AI, employees learn how to identify AI improvement opportunities.&nbsp;</p><h3>7. Create an AI Sandbox for Open Source&nbsp;</h3><p class="">Besides specific workshops, it is also valuable to create dedicated spaces for employees to be able to work with artificial intelligence.&nbsp; The open-source community is alive with AI-related capabilities – models, code libraries, and more (e.g., <a href="https://deepmind.google/technologies/alphafold/" target="_blank">AlphaFold</a>) – that can be used to learn about and explore AI-powered improvements.&nbsp; However, due to the public nature of these assets, risks related to security, quality, service level disruption, and compliance need to be contained.&nbsp; And since many AI solutions need access to company data in some form, many organizations prefer to host these technologies in environments where they have more direct control of leveraged data sets.&nbsp;</p><p class="">Cloud and “virtual machine” technologies offer great mechanisms for standing up short-term “sandboxes” for exploring new AI projects.&nbsp; Most major cloud hosting companies have semi-automated procedures for quickly standing up and configuring new environments.&nbsp; Code access is restricted to information explicitly provided in the sandbox, and external communications can be configured to prevent disclosure of company information to third parties.&nbsp; Once the exploration is done, shutting the environment down is almost as easy as pressing the delete key.&nbsp;</p><h3>8. Launch a Pilot Project&nbsp;</h3><p class="">As life sciences organizations explore the world of possibilities with artificial intelligence, eventually at least one idea will surface that holds promise.&nbsp; Though AI opportunities may look like IT projects, the greatest impact will be felt when leaders treat AI investments as more comprehensive business initiatives.&nbsp;</p><p class="">What does that mean?&nbsp; Consider the following criteria in spinning up an AI pilot:&nbsp;</p><ul data-rte-list="default"><li><p class=""><strong>Strategy &amp; Objectives:</strong> are the goals clear and measurable?&nbsp; Does everyone understand how the goals align to the company’s strategy?&nbsp;</p></li><li><p class=""><strong>Scope &amp; Resourcing:</strong> is the project sized for success (i.e., specific, focused, impactful)? Do we have the right business and technical experts engaged?&nbsp;</p></li><li><p class=""><strong>Use Case and Success Criteria:</strong> is the workflow (both “as-is” and “to-be”) well defined? Does everyone agree on what a great outcome looks like, and what measurable value is being generated?&nbsp; Do we know when we are done with this project?&nbsp;</p></li><li><p class=""><strong>Architecture &amp; Data Strategy:</strong> have we considered how the design of this AI capability might fit within our enterprise architecture (e.g., integration, authentication, authorization)?&nbsp; Do we know how this solution would scale up over time? Do we have a plan for providing the AI high-quality data?&nbsp; Are we clear on how to address privacy, regulatory, or data use constraints?&nbsp;</p></li><li><p class=""><strong>Governance &amp; Communication:</strong> are executive sponsorship and decision-making accountabilities clear?&nbsp; Are all stakeholders on board and willing to implement changes if the project is successful? Have we agreed on how, when, and where to share information and updates about the project?&nbsp;</p></li><li><p class=""><strong>Quality &amp; Compliance:</strong> do we know how we will assess the quality of the work we produce, the data we provide, and the results generated by AI?&nbsp; Are testing, validation, and user acceptance requirements and procedures clear?&nbsp;</p></li><li><p class=""><strong>Deployment &amp; Change Management:</strong> do we know how we intend to deploy, transition, and train?&nbsp; What are our plans for documentation, maintenance, support, and continuous improvement?&nbsp;</p></li><li><p class=""><strong>Risks &amp; Contingencies:</strong> have we agreed on how to assess and manage risk?&nbsp; Do we know what to do if the solution does not work as intended?&nbsp; What, if any, ethical concerns need to be managed?&nbsp;</p></li></ul><p class="">Of course, you don’t need the answer to all those questions to get started with a pilot.&nbsp; But ensuring leaders and pilot teams are thinking holistically about the program will help to ensure that successful pilots can be converted to successful new business capabilities.&nbsp;</p><p class="">If you’d like to learn more about AI technologies in life sciences, consider reading my <a href="https://creoconsulting.com/campaign/ten-ai-trends-every-life-sciences-leader-should-know/" target="_blank">white paper</a>.&nbsp; It provides an overview of some of the key trends life sciences leaders need to be aware of when considering AI investments.&nbsp; And <a href="https://www.linkedin.com/in/jasonburke/" target="_blank">reach out to me directly</a> if you’d like to talk more about any of these concepts, I’m always curious to learn how others are thinking about this exciting space.</p>]]></content:encoded><media:content height="752" isDefault="true" medium="image" type="image/jpeg" url="https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/1733072936862-WO2KC75UOAIO1ZIFFKDF/AI_keyboard_user.jpeg?format=1500w" width="1500"><media:title type="plain">Eight Ways Life Sciences Leaders Can Kickstart Artificial Intelligence</media:title></media:content></item><item><title>Unlocking Data Governance for Strategic Growth</title><category>Data</category><dc:creator>Jason Burke</dc:creator><pubDate>Thu, 16 May 2024 15:11:00 +0000</pubDate><link>https://www.jasonburke.online/blog/data-governance-for-strategic-growth</link><guid isPermaLink="false">674a65c6ba0c721888a63309:674a6885532ea52b2bdb097c:674a6885532ea52b2bdb0986</guid><description><![CDATA[Though data governance is often a taboo topic among executives, thoughtful 
investments in data asset improvements can have a transformative impact on 
data trust and AI capabilities.]]></description><content:encoded><![CDATA[<p class="">In today’s digital era, organizations face a rapidly growing need to unlock value from their data. From enabling artificial intelligence (AI) to powering data-driven decision-making, the path to achieving these transformative goals often begins with strong data governance.</p><p class="">For many health and life sciences businesses, however, the opportunity to fully harness  data is hindered by gaps in their IT and data strategies. This gap reflects a broader challenge: despite significant investments in technology, many organizations struggle to ensure their data is accurate, accessible, and trustworthy. Building robust data governance capabilities is essential for turning these investments into actionable insights and sustained value.</p><h3>Why Data Governance Matters</h3><p class="">Data governance is the cornerstone of data utility and trust. It establishes the processes, policies, and accountability frameworks necessary to manage data assets effectively. Without it, even the most advanced analytics or AI initiatives can falter due to poor data quality, inconsistent definitions, and insufficient data availability.</p><p class="">Data governance is also a term that scares many leaders.  The topic has a mixed history of  implementations, with many companies reporting that their data governance programs failed to produce meaningful impacts.  I frequently find myself trying to help executives understand their myriad of data problems without using the term “data governance” because the topic is nearly taboo.</p><h3>Typical Misconceptions with Data Governance</h3><p class="">Organizations often encounter several recurring barriers to implementing effective data governance. These challenges can stall progress, leaving them unable to capitalize on their data’s potential. Below are seven common misconceptions that hinder progress and ways to overcome them:</p><p class="">	1.	<strong>“We Lack the Right Technology”</strong></p><p class="">The perception that data issues are primarily technological is widespread but often misplaced. Many improvements in data governance stem from better processes, roles, and accountability—not from expensive new tools.</p><p class="">	2.	<strong>“It’s Too Time-Consuming”</strong></p><p class="">Governance initiatives do require time and effort, but they also create efficiencies that save time in the long run. Teams empowered to fix systemic data issues will spend less time firefighting repetitive problems.</p><p class="">	3.	<strong>“The Problems Are Too Complex or Expensive to Solve”</strong></p><p class="">Data governance is often seen as daunting. However, simple, targeted initiatives—such as cleaning up high-impact datasets or addressing siloed data ownership—can yield significant improvements.</p><p class="">	4.	<strong>“We’ll Fix Issues on Demand”</strong></p><p class="">Tackling data quality reactively leads to growing technical debt and project inefficiencies. Proactive governance is not only cost-effective but also scalable.</p><p class="">	5.	<strong>“We Don’t Have the Right Expertise”</strong></p><p class="">While data governance expertise is specialized, resources are increasingly available through training, frameworks, and consulting services. Building a foundation of knowledge internally is more achievable than ever.</p><p class="">	6.	<strong>“We’ll Start After a Key Event”</strong></p><p class="">Waiting for events such as mergers or AI deployments to begin data governance often backfires. Governance should precede these milestones to ensure they proceed efficiently and with reduced risk.</p><p class="">	7.	<strong>“There’s Too Much Political Risk”</strong></p><p class="">Effective data governance can depoliticize challenges, fostering collaboration across silos. Transparent, well-communicated goals help align stakeholders and mitigate resistance.</p><h3>Bridging the Gap Between Business and IT</h3><p class="">Effective data governance requires more than just IT involvement; it depends on strong partnerships between business and IT teams. While IT often recognizes the technical requirements for governance, business leaders are closer to the operational pain points of poor data quality. Aligning these perspectives ensures governance efforts address both strategic goals and day-to-day operational needs.</p><h3>Taking the First Steps</h3><p class="">A common pitfall is attempting to launch a top-down, enterprise-wide initiative without first demonstrating value. Instead, consider starting with smaller, impactful projects that focus on a single business process or department. For instance:</p><p class="">	•	<strong>Customer onboarding:</strong> Mapping data quality issues tied to customer journeys can deliver immediate benefits to multiple stakeholders.</p><p class="">	•	<strong>Operational reporting:</strong> Cleaning and standardizing data for key metrics can build trust in decision-making processes.</p><p class="">Once early wins are established, these successes can be used to scale efforts, fostering champions across the organization and increasing executive buy-in.</p><h3>Enabling Data-Driven Transformation</h3><p class="">Robust data governance does more than ensure clean, consistent data—it lays the foundation for transformational capabilities like AI. AI initiatives thrive on well-curated, high-quality data that is timely, reliable, and accessible. By prioritizing governance, organizations can move beyond tactical fixes toward scalable, strategic insights.</p><p class="">As the business landscape grows increasingly reliant on data, the organizations that succeed will be those that take governance seriously—not as a checkbox exercise, but as a strategic imperative. By addressing barriers and taking meaningful first steps, executives can position their organizations to leverage their data as a critical asset for sustained growth and innovation.</p>]]></content:encoded><media:content height="1057" isDefault="true" medium="image" type="image/jpeg" url="https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/1733075918018-WL7JUBV2QBMWIP504TLF/adam-nowakowski-D4LDw5eXhgg-unsplash.jpg?format=1500w" width="1409"><media:title type="plain">Unlocking Data Governance for Strategic Growth</media:title></media:content></item><item><title>FDA IT Strategy in a World of AI</title><category>Regulatory</category><dc:creator>Jason Burke</dc:creator><pubDate>Thu, 09 Nov 2023 16:12:00 +0000</pubDate><link>https://www.jasonburke.online/blog/blog-post-title-three-34gpy</link><guid isPermaLink="false">674a65c6ba0c721888a63309:674a6885532ea52b2bdb097c:674a6885532ea52b2bdb0984</guid><description><![CDATA[The US Food and Drug Administration’s updated IT strategy highlights the 
agency’s planned direction and investments while also demonstrating 
fundamental industry challenges ahead.]]></description><content:encoded><![CDATA[<p class="">A few weeks ago, the US Food and Drug Administration&nbsp;<a href="https://www.regulations.gov/document/FDA-2023-N-3636-0001" target="_blank">announced</a>&nbsp;the availability of an updated&nbsp;<a href="https://www.fda.gov/media/172067/download" target="_blank">FDA IT Strategy</a>&nbsp;for comment.&nbsp; The document, which builds on the agency’s broader&nbsp;<a href="https://www.fda.gov/about-fda/office-digital-transformation/odt-reports" target="_blank">modernization initiatives</a>, is an enterprise plan that is intended to “guide FDA’s Information Technology (IT) direction and investments for the next four years.”&nbsp; And it highlights some of the fundamental challenges ahead for&nbsp;<a href="https://www.creoinc.net/it-digital-transformation/" target="_blank">life sciences digital transformation</a>.</p><h3>The FDA IT Strategy</h3><p class="">For anyone operating in the life sciences area, the FDA IT Strategy is an interesting read.&nbsp; The strategy outlines six top-level goals in developing a more enterprise-focused technology strategy, which can be roughly summarized as:</p><ol data-rte-list="default"><li><p class="">Creating a shared technology ecosystem across the organization designed to support communication, collaboration, trusted data, and transparency.</p></li><li><p class="">Improving the base infrastructure through cloud technologies, flexible service orientation, and zero-trust principles.</p></li><li><p class="">Cultivating scalable, cost-effective, tailorable, and secure enterprise services through stronger business alignment, digital transformation initiatives, and customer experiences.</p></li><li><p class="">Embracing data as a core asset, including data governance, literacy, accessibility, exchange, and advanced analytics.</p></li><li><p class="">Pursuing balanced, responsible, AI-based innovations that align to the organization’s mission.</p></li><li><p class="">Attracting, cultivating, and retaining the right leadership and workforce.</p></li></ol><p class="">At first glance, these seem like the right set of priorities.&nbsp; In fact, it would be difficult to debate most of them: would anyone argue for less modernized technology that ignores data, for example?&nbsp; The list is certainly consistent with non-agency IT priorities – virtually all of these strategic goals could have been copied from the business and technology plans of the agency’s life sciences stakeholders. &nbsp;Regardless of organizational size or mission, these technology forces are top of mind for many industry leaders today.</p><p class="">The FDA IT Strategy has several recurring themes that are worth calling out.&nbsp; First, there is a strong recognition of the need for stakeholder engagement and alignment.&nbsp; That point is particularly important for the second recurring theme, which is around business process and service scalability.&nbsp; If the agency is reflecting the broader IT ecosystem, leaders see process orchestration, automation, and streamlining as critical enablers for capturing economies of scale.&nbsp; And, of course, a third recurring theme is unleashing the opportunity in AI, analytics, and data.</p><h3>Artificial Intelligence and the Evolving Policy Landscape</h3><p class="">Acknowledging the logic of the proposed plan, perhaps the largest issue facing the agency today will be more cultural than technological: how do you make an agency with the size, complexity, and longevity of the FDA agile? &nbsp;That question has been asked before (i.e., combinatorial chemistry, precision therapy development), though it seems clear to most industry insiders that the pace of innovation in areas like AI is far greater than the ability of any federal agency to respond effectively.</p><p class="">But that’s not for a lack of trying.&nbsp; The fall of 2023 has seen a flurry of policy and regulatory activity related to AI.&nbsp; On October 30, 2023, President Biden issued an Executive Order on&nbsp;<a href="https://www.whitehouse.gov/briefing-room/statements-releases/2023/10/30/fact-sheet-president-biden-issues-executive-order-on-safe-secure-and-trustworthy-artificial-intelligence/" target="_blank">Safe, Secure, and Trustworthy Artificial Intelligence</a>, the most significant executive branch foray to date in introducing AI safety and security measures into public policy. The proposed actions are&nbsp;<a href="https://www.technologyreview.com/2023/10/30/1082678/three-things-to-know-about-the-white-houses-executive-order-on-ai" target="_blank">primarily focused</a>&nbsp;on mitigating the rapidly emerging risks associated with AI, addressing issues such as transparency, labelling, testing, and standards development.&nbsp; From a life sciences perspective, the order was notable for calling out issues such as stronger provisions against algorithmic discrimination; protections for the engineering of dangerous biological materials; promoting AI research in healthcare; strengthening consumer privacy protections; and ensuring the safe and responsible use and reporting of AI in healthcare and drug development.</p><p class="">That White House Executive Order was followed two days later by the Office of Management and Budget’s&nbsp;<a href="https://www.whitehouse.gov/omb/briefing-room/2023/11/01/omb-releases-implementation-guidance-following-president-bidens-executive-order-on-artificial-intelligence/" target="_blank">release</a>&nbsp;of a new&nbsp;<a href="https://www.whitehouse.gov/wp-content/uploads/2023/11/AI-in-Government-Memo-draft-for-public-review.pdf" target="_blank">draft policy</a>&nbsp;on Advancing Governance, Innovation, and Risk Management for Agency Use of Artificial Intelligence.&nbsp; The OMB policy, which is open for comment through December 5th, 2023, is intended to offer richer guidance to federal agencies regarding AI strategy development, governance, responsible use, transparency, reporting, protections, and risk management.&nbsp; In parallel, NIST continues to advance their&nbsp;<a href="https://www.nist.gov/itl/ai-risk-management-framework" target="_blank">AI Risk Management Framework</a>, and all of these activities are advancing in the context of the White House’s Blueprint for an&nbsp;<a href="https://www.whitehouse.gov/ostp/ai-bill-of-rights/" target="_blank">AI Bill of Rights</a>.</p><p class="">This oversight momentum is not specific to the US.&nbsp; The European Union’s proposed&nbsp;<a href="https://artificialintelligenceact.eu/" target="_blank">AI Act</a>&nbsp;offers a tiered risk model for AI-related regulatory obligations, and has already garnered&nbsp;<a href="https://www.usnews.com/news/technology/articles/2023-11-08/eu-ai-act-to-serve-as-blueprint-for-global-rules-benifei-says" target="_blank">significant traction</a>.&nbsp; In addition, the Group of Seven industrial nations produced an 11-point&nbsp;<a href="https://www.reuters.com/technology/g7-agree-ai-code-conduct-companies-g7-document-2023-10-29/" target="_blank">voluntary code of conduct</a>&nbsp;regarding the use of AI. The code “aims to promote safe, secure, and trustworthy AI worldwide and will provide voluntary guidance for actions by organizations developing the most advanced AI systems, including the most advanced foundation models and generative AI systems.”&nbsp; Interestingly, the stronger policies surfacing in western regulatory environments are&nbsp;<a href="https://www.reuters.com/technology/southeast-asia-eyes-hands-off-ai-rules-defying-eu-ambitions-2023-10-11/" target="_blank">not consistently mirrored in some Asian territories</a>&nbsp;where countries are adopting a more relaxed posture.</p><h3>Building a Strong IT Strategy</h3><p class="">Forces like these place new obligations on the FDA to adopt technology strategies that strike the right balance between innovation, standardization, safety, quality, and scientific rigor.&nbsp; It’s a tall order to be sure. Alongside other federal agencies, DHHS has provided a&nbsp;<a href="https://www.hhs.gov/sites/default/files/hhs-trustworthy-ai-playbook.pdf" target="_blank">Trustworthy AI Playbook</a>&nbsp;alongside an&nbsp;<a href="https://ai.gov/ai-use-cases/" target="_blank">inventory of over 160 existing AI use cases</a>, over 40 of which are&nbsp;<a href="https://www.hhs.gov/sites/default/files/hhs-ai-use-cases-2023-public-inventory.csv" target="_blank">specific to the FDA</a>.&nbsp; The question for FDA technology leaders – like their non-agency counterparts in the industry – is how to develop the right data and analytical competencies that can support the rising tide of AI and data sciences innovations and associated medical advancements.</p><p class="">The industry is clearly interested in the answer to that question.&nbsp; The request for comments elicited feedback from a variety of organizations including Google, Health Verity, McKesson, The RWE Alliance, and others.&nbsp; The nature of the comments varied widely, ranging from poorly masked advertisements to deeper dives into important technology considerations like APIs and data standards.</p><p class="">One of the more notable responses came in the form of a combined&nbsp;<a href="https://www.regulations.gov/comment/FDA-2023-N-3636-0014" target="_blank">letter from BIO and PhRMA</a>&nbsp;which echoed some of my own reflections on the document.&nbsp; As a strategist looking at the FDA IT Strategy document, there were five areas where I felt the strategy could be improved, especially given the needs for AI-related agility and the themes above:</p><ol data-rte-list="default"><li><p class=""><strong>Connect explicitly to business goals.</strong>&nbsp; The best IT strategies directly connect their goals and tactics to corresponding business objectives.&nbsp; By illustrating what IT investments are supporting which business goals, leaders can discern whether planned efforts are likely to deliver the needed results.</p></li><li><p class=""><strong>Clearly define current and future states.</strong>&nbsp; BIO and PhRMA shared that the strategy as stipulated reads more like a list of principles, and I think that’s right.&nbsp; Effective strategies are based on an unambiguous, detailed description of both the current and future states of the organization.&nbsp; What specifically are you committing to build, improve, re-engineer, etc.? &nbsp;Those more detailed commitments should deliver business value that can be measured, and they offer opportunities for cross-functional (and cross-industry) alignment. This is especially important in the space of AI, where the deafening roar of hype can easily obscure the real opportunities for impact.</p></li><li><p class=""><strong>Make Goals Empirical.</strong>&nbsp; Strategies can often read as a list of attractive aspirations – sometimes called “motherhood and apple pie” – that don’t seem actionable.&nbsp; When strategic goals have measurable performance metrics associated with each goal, it becomes easier to align investments and tactics to drive their growth.&nbsp; How do we know if our IT strategy is being successful in year 1?&nbsp; In year 2?&nbsp; In year 3?&nbsp; What is the benchmark from which we are measuring improvements?&nbsp; These questions might seem cumbersome in their detail, but they actually help focus attention on the most value-creating activities.</p></li><li><p class=""><strong>Include more “how.”</strong>&nbsp; A good strategy helps you know when to say “no” to the continually emerging landscape of attractive opportunities that risk diluting resources. &nbsp;We don’t just want to know what is “good”; we also want to know what is “important.” By intentionally selecting how a strategy is being pursued, organizations can more easily identify what activities and investments align with a focused execution of the strategy. For example, much like an electric car needs batteries, a comprehensive and effective data strategy would seem to be a critical element of the FDA’s growth plans. We often frame these elements of a strategy as strategic imperatives: the areas of operational focus and excellence required in order for a strategy to be successful. The FDA’s plan covers the need to share, govern, and improve the accessibility of data. But it does not reflect an explicit priority for a data strategy and architecture that directly address complex industry business process use cases such as cross-study data analysis; rapidly harnessing existing data for AI model training; near real-time industry data sharing with the agency; and other improvements in electronic information structures and submissions. These potential process automations can improve the timeliness of regulatory reviews and approvals. A more detailed version of the agency’s own&nbsp;Data Modernization Action Plan&nbsp;could potentially address these issues in the future, creating an opportunity for life sciences stakeholders to align their enterprise architecture plans and standards to better support the agency’s operating model.</p></li><li><p class=""><strong>Incorporate timetables.</strong>&nbsp; Strategies are usually not executed in weeks or months. &nbsp;Typically, there should be a logical progression – a roadmap of milestones, phases, or success measures – that establish the expected implementation of strategic plans and associated improvements in performance.&nbsp; In fairness, the strategy document refers to a future deliverable that will offer more of an IT roadmap, but ideally such phasing would also be reflected at this level of planning granularity as well.</p></li></ol>


  




  




  
  <p class="">Interestingly, much like the strategy itself, these improvement opportunities also commonly show up in the life sciences stakeholder organizations – so much so that part of the&nbsp;ISEP strategic planning methodology&nbsp;we use is designed to proactively address them.&nbsp; And to be sure, the FDA’s CIO Vid Desai and the entire team at the agency should be commended on building this strategy.&nbsp; It’s advancing an important conversation that can directly impact the speed at which patients around the world receive tomorrow’s life-saving therapies.&nbsp; I’ll be excited to review the next iteration.</p>]]></content:encoded><media:content height="729" isDefault="true" medium="image" type="image/jpeg" url="https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/1733074804487-TTOQNXOZFXG7O2IQ3AGJ/AI_keyboard.jpg?format=1500w" width="1166"><media:title type="plain">FDA IT Strategy in a World of AI</media:title></media:content></item><item><title>Managing Emerging AI Cybersecurity Risks</title><category>security</category><dc:creator>Jason Burke</dc:creator><pubDate>Wed, 05 Jul 2023 18:29:00 +0000</pubDate><link>https://www.jasonburke.online/blog/managing-ai-cybersecurity-risks</link><guid isPermaLink="false">674a65c6ba0c721888a63309:674a6885532ea52b2bdb097c:674e0a9737f5e2577fced504</guid><description><![CDATA[Though offering impressive capabilities, the rapid adoption of artificial 
intelligence capabilities within the life sciences and healthcare sectors 
also brings notable risks that must be managed.]]></description><content:encoded><![CDATA[<p class="">Every cybersecurity professional faces the challenge of managing cybersecurity risks associated with emerging technologies such as artificial intelligence (AI).&nbsp; Traditionally, most experts agree even the best cybersecurity solutions are only one step ahead of (or even sometimes behind) the capabilities of attackers.&nbsp; This situation is especially difficult when the pace of the emerging technology adoption is high.</p><p class="">Life sciences and healthcare are embracing AI as a fundamental tenet of&nbsp;<a href="https://www.creoinc.net/how-ai-is-reinventing-medical-innovation/" target="_blank">medical innovation</a>.&nbsp; And the past year has produced an unprecedented level of interest in AI applications, especially “generative AI” techniques found in solutions such as&nbsp;<a href="https://web.archive.org/web/20240501142450/https://openai.com/blog/chatgpt/" target="_blank">ChatGPT</a>,&nbsp;<a href="https://github.com/features/copilot" target="_blank">GitHub Copilot</a>,&nbsp;<a href="https://bard.google.com/" target="_blank">Bard</a>,&nbsp;<a href="https://openai.com/index/dall-e-2/" target="_blank">DALL-E 2</a>,&nbsp;<a href="https://www.midjourney.com/" target="_blank">Midjourney</a>, and other offerings.&nbsp; Due to the sophisticated functionality and broad availability of these products, they are having a disruptive effect on many life sciences organizations, introducing a wide array of risks that bad actors can exploit.</p><h3>Nine AI-related Cybersecurity Risks</h3><p class="">Predictive analytics, including artificial intelligence and machine learning methods, are far from new.&nbsp; Given the life-critical nature of analytical and decision support applications within life sciences and healthcare, AI risk analysis in these industry settings often focuses on the AI model’s data and resulting performance (e.g., biased or erroneous data, ethical considerations, incorrect and opaque algorithms, improperly tuned models,&nbsp;<a href="https://en.wikipedia.org/wiki/Hallucination_(artificial_intelligence)" target="_blank">AI hallucinations</a>).</p><p class="">For cybersecurity and risk management professionals, many emerging risks are also related to how users – both employees and external stakeholders – are adopting and exploiting these new technology capabilities. &nbsp;If you think the risks are not that notable, consider that Apple, Samsung, Amazon, Verizon, Citigroup, Goldman Sachs, Wells Fargo, and Accenture among others have already&nbsp;<a href="https://www.hcamag.com/ca/specialization/employment-law/several-employers-ban-chatgpt-use-by-employees/437646" target="_blank">restricted the use</a>&nbsp;of these technologies.</p><p class="">What risks are they concerned about?&nbsp; Broadly speaking, AI-related risks commonly fall into nine categories:</p><ol data-rte-list="default"><li><p class=""><strong>Loss of Intellectual Property</strong>.&nbsp; In a&nbsp;<a href="https://www.prweb.com/releases/nearly_half_of_senior_leaders_believe_colleagues_have_inadvertently_shared_corporate_data_with_chatgpt/prweb19332734.htm" target="_blank">recent survey</a>, approximately half of the senior executive respondents believe their company data has been inadvertently shared with ChatGPT.&nbsp; Employees using generative AI tools – especially those hosted outside their corporate infrastructure – may provide copies of corporate data as required inputs to the AI system without realizing that they are disclosing confidential information to an unauthorized vendor.&nbsp; Such information is then subject to further disclosures through system hacking, internal product vulnerabilities, or system use by competitors.&nbsp; And when such information is governed by regulations, the risks increase further.</p></li><li><p class=""><strong>Contractual and regulatory violations.</strong>&nbsp; Disclosure of some information to AI systems may also constitute violations of contractual terms or regulatory obligations. For example, most companies have non-disclosure agreements in place with their data providers.&nbsp; And data critical for the development and training of many industry-specific AI algorithms – patient data, for example – is protected by HIPAA which restricts both the use and disclosure of that data.&nbsp; This issue is particularly sensitive when dealing with any data that carries privacy concerns.&nbsp;</p></li><li><p class=""><strong>Social engineering.</strong>&nbsp; Generative AI tools can be leveraged by attackers to engage in social engineering attacks, tricking users into revealing sensitive information or performing malicious actions. Because the AI models can conjure highly realistic imitations of individual leaders and employees within an organization, these phishing attacks are harder to detect by the average user.&nbsp; Attackers are increasingly running their email campaigns through natural language models to eliminate broken English and misspellings which have been a tell-tale sign of phishing attempts. &nbsp;And with a few recorded minutes of an individual’s voice, AI voice generators can recreate any voice message desired (e.g., this is the CEO and I need a favor urgently for this afternoon’s board meeting).</p></li><li><p class=""><strong>Copyright infringement.</strong>&nbsp; Generative AI models are trained on existing information works that are usually copyrighted.&nbsp; Currently, the boundaries concerning the appropriate use of copyrighted works as a foundation for new works are unclear.&nbsp; If a computer generates a new work by mimicking an existing work, is the new work infringing on that existing work’s copyright?&nbsp; Like all things related to copyright law, these issues are expected to take considerable time to normalize.</p></li><li><p class=""><strong>Larger pool of stronger adversaries.</strong>&nbsp; The broad and open availability of sophisticated AI tools creates more opportunities for new bad actors (e.g., hackers, malicious content creators) and software threats (e.g., malware, bots) to emerge.&nbsp; Individual hackers, groups, extremists, and geopolitical / nation-state actors have more tools in their toolbox today.&nbsp; These tools can increase the effectiveness of their attacks (e.g., AI-driven vulnerability and attack planning), and are increasingly easy to use by novices.</p></li><li><p class=""><strong>Code vulnerabilities.</strong>&nbsp; AI can also be used to analyze and identify vulnerabilities in 3rd&nbsp;party software code in web applications, websites, and network configurations.&nbsp; This means that attackers can find and exploit zero-day vulnerabilities much faster.&nbsp; In addition, most of the AI community leverages open-source software development tools and code libraries for developing new AI capabilities.&nbsp; The continuously evolving state of these systems and code bases creates a potential exposure point for enterprises that introduce them into their infrastructure (e.g., software defects, malicious code insertion).</p></li><li><p class=""><strong>Newer forms of viruses and malware.</strong>&nbsp; Any machine infected with malicious code is obviously a threat.&nbsp; As viruses and malware increasingly incorporate AI capabilities, these threats will become increasingly sophisticated, able to learn from and adapt to behaviors and new experiences.&nbsp; For example, malicious code can leverage AI to change its digital fingerprint, thereby enabling it to evade detection by security software.</p></li><li><p class=""><strong>Misinformation campaigns.</strong>&nbsp; The ability of generative AI to create&nbsp;<a href="https://mashable.com/article/moments-ai-fooled-internet-deepfakes-misinformation" target="_blank">realistic imitations</a>&nbsp;is not limited to social engineering hacks.&nbsp; Competitors, activists, and disgruntled employees now have access to tools that can fabric audio and video assets that can damage an organization’s reputation and market valuation.&nbsp; Though popular concerns about this issue have largely focused on&nbsp;<a href="https://www.washingtonpost.com/politics/2023/06/22/cyber-argument-regulating-ai/" target="_blank">political and regulatory issues</a>, the risks are real for any business, organization, or individual.</p></li><li><p class=""><strong>Device vulnerabilities.</strong>&nbsp; Specialized computing hardware, medical devices, and other solutions that include embedded AI algorithms need to be effectively managed over their lifecycle.&nbsp; AI models need to be periodically tuned.&nbsp; Operating systems and firmware need to be hardened.&nbsp; Application software and interfaces needs to be patched.&nbsp; Organizations that deploy these assets without having a strong lifecycle management process in place introduce vulnerabilities to their operational infrastructure as these technologies age.</p></li></ol><h3>Addressing AI-related Cybersecurity Risks</h3><p class="">If the list of risks above seems daunting, it need not be.&nbsp; There are proven solutions and practices that help organizations address those concerns. &nbsp;Key components of an effective cybersecurity approach for managing AI-related risks include the following:</p><ol data-rte-list="default"><li><p class=""><strong>Cybersecurity Strategy</strong>: Do you have a strong security and privacy program today?&nbsp; Are your threat operations positioned to respond to emerging risks?&nbsp; For many organizations, increasing their readiness for AI starts with increasing their&nbsp;<a href="https://www.creoinc.net/cybersecurity/" target="_blank">cybersecurity posture</a>&nbsp;overall.&nbsp; A well-formed cybersecurity strategy provides a solid framework for securing assets, managing operations, and responding to threats.</p></li><li><p class=""><strong>Enterprise Architecture</strong>: Every life sciences organization needs a&nbsp;<a href="https://www.creoinc.net/life-sciences-it-strategy/" target="_blank">strong IT strategy</a>&nbsp;and associated enterprise architecture.&nbsp; Well-managed enterprise architectures provide a wealth of cybersecurity benefits – minimizing security perimeters, maximizing protective measures, standardizing security methods, and more – that serve to make the organization’s infrastructure more resilient to attacks.</p></li><li><p class=""><strong>Data Mapping &amp; Governance</strong>: What are your organization’s high-value data assets, and how are they protected?&nbsp; How operationally do you mitigate the risks of accidental disclosure?&nbsp; The answers to these critical questions vary widely by company, and effective answers require a deep understanding of each organization’s mission, products, operating model, customers, and regulatory requirements.</p></li><li><p class=""><strong>Risk Management</strong>:&nbsp; Countering AI cyber threats requires smart risk management.&nbsp; The landscape of risk is changing each week as AI innovations continue to advance.&nbsp; Until cybersecurity technologies can catch up, the best defense continues to be good risk management practices.</p></li><li><p class=""><strong>User Awareness &amp; Training:</strong>&nbsp; The first, last, and best line of defense for AI-related risks is educated users.&nbsp; Many leaders and front-line users are simply not aware of the risks associated with using generative AI tools and other emerging technologies. Though traditional cybersecurity awareness training does not cover AI, organizations like&nbsp;<a href="https://www.creoinc.net/cybersecurity/" target="_blank">CREO</a>&nbsp;offer specific training covering new attack vectors presented by AI and how users can help the organization manage the evolving new threats.</p></li></ol><p class="">Finally, it is also worth remembering that AI is not just a source of risks – it’s also a highly valuable defense tool.&nbsp; A modernized arsenal of cybersecurity technologies includes AI-powered capabilities in real-time risk assessment and quantification; real-world security telemetry, detection, and response; self-healing systems; endpoint intelligence; zero-trust asset management; and intelligent vulnerability prioritization and management.&nbsp; When coupled with the practices above, life sciences and healthcare organizations can control the risks introduced by AI while also reaping the benefits of these powerful technologies.</p>]]></content:encoded><media:content height="1000" isDefault="true" medium="image" type="image/jpeg" url="https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/1733168049158-JRVUYS8UU6KKWJ2NWBZ4/unsplash-image-iIJrUoeRoCQ.jpg?format=1500w" width="1500"><media:title type="plain">Managing Emerging AI Cybersecurity Risks</media:title></media:content></item><item><title>Building Data-Driven Cultures Through Trust</title><category>Data</category><dc:creator>Jason Burke</dc:creator><pubDate>Mon, 21 Nov 2022 19:39:00 +0000</pubDate><link>https://www.jasonburke.online/blog/building-data-driven-cultures-through-trust</link><guid isPermaLink="false">674a65c6ba0c721888a63309:674a6885532ea52b2bdb097c:674e0cfcb381bd05024992e1</guid><description><![CDATA[Organizational aspirations and plans for becoming more “data driven” often 
do not reflect the basic reality that building a data-driven culture is a 
change management challenge rooted in trust.]]></description><content:encoded><![CDATA[<h3>Is your company driven by data? </h3><p class="">If so, you may be in the minority. Data from multiple surveys of Fortune 1000 executives show that only around 20-25% companies report successfully establishing a data-driven culture.</p><p class="">The lack of progress is not a reflection of a lack of investment. New leadership roles – Chief Analytics Officer, Chief Data Officer – have emerged to facilitate value creation through better insights. A lot of industry attention is being directed on data literacy and self-service enablement as key to unlocking a more data-driven workforce. And company executives are frequently communicating to their investors that their strategies and investments reflect data- and insight-driven opportunities (e.g., AI).</p><h3>So where is the gap?</h3><p class="">Though corporate priorities may reflect data-oriented offerings and business processes, organizational plans often do not reflect the basic reality that building a data-driven culture is a change management challenge. Senior leaders need to agree on (and appropriately resource) an effective approach for transforming the culture, helping employees transition their activities and responsibilities from the “as-is” state to “to-be” state.</p><p class="">For example, is the organization approaching this transition through a specific program that addresses&nbsp;<a href="https://en.wikipedia.org/wiki/John_Kotter" target="_blank">Kotter’s change management</a>&nbsp;components (e.g., creating urgency, building change teams, maintaining change progress, etc.)? Or is the organization pursuing a more iterative, opportunistic approach such as&nbsp;<a href="https://en.wikipedia.org/wiki/Nudge_theory" target="_blank">nudge theory</a>? The approach can be big or small, but being intentional about change management enables leaders to adequately address the most common barriers to adoption.</p><p class="">One way to think about the change management dimension is through the lens of trust. Trust is the cornerstone of any successful change management program. So what would it look like for an employee to trust a new data-driven insight enough to confidently change their behavior and/or mindset? In senior executive education courses I sometimes teach, we explore six areas where trust contributes to the development of data-driven cultures:</p><ul data-rte-list="default"><li><p class=""><strong>Trust with Leadership</strong>&nbsp;– does this insight have executive attention and support? Employees tend to embrace insights that their leaders visibly value (i.e., model as important and safe). In a sense, there is a transfer of trust.</p></li><li><p class=""><strong>Trust with Impact&nbsp;</strong>– does this insight matter? When insights are strongly aligned to actionable decisions and behaviors, they more reliably produce trustworthy results.</p></li><li><p class=""><strong>Trust with People</strong>&nbsp;– is the insight driven by experts? Employees want to know that the people who best understand the problem space contributed to the insight and interpretations.</p></li><li><p class=""><strong>Trust with Data</strong>&nbsp;– is this insight based on good information? If employees don’t have confidence in the underlying data source, they are much less likely to trust and embrace the insight.</p></li><li><p class=""><strong>Trust with Development</strong>&nbsp;– are the data and its insights created using best practices? Employees know that data can sometimes be misleading; they want confidence that the process used to create the insight was well designed and well executed.</p></li><li><p class=""><strong>Trust with Communications</strong>&nbsp;– is the insight clearly visible and understood? Employees cannot embrace an insight they don’t see or cannot interpret.</p></li></ul><p class="">A few of these areas tend to receive more “airtime” in discussions than others; for example, the majority of executives do not trust their data today. The irony is that shortcomings within one area such as “trust in data” is often a downstream effect of deficits in other areas such as “trust in development.”</p><p class="">Where in your organization do you see problems with trust inhibiting the development of a more data-driven business and culture?</p>]]></content:encoded><media:content height="1054" isDefault="true" medium="image" type="image/jpeg" url="https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/1733168557800-YECSQFY6O6RUXK3HO9GZ/unsplash-image-ZoVR7mPHMGo.jpg?format=1500w" width="1500"><media:title type="plain">Building Data-Driven Cultures Through Trust</media:title></media:content></item><item><title>How AI is Reinventing Medical Innovation</title><category>AI</category><dc:creator>Jason Burke</dc:creator><pubDate>Thu, 14 Jul 2022 18:49:00 +0000</pubDate><link>https://www.jasonburke.online/blog/ai-is-reinventing-medical-innovation</link><guid isPermaLink="false">674a65c6ba0c721888a63309:674a6885532ea52b2bdb097c:674e0f46f10cab46b7bf47ba</guid><description><![CDATA[The health and life sciences industries are facing a paradigm shift where 
emerging technologies like artificial intelligence (AI) and genomic 
sciences are unlocking new ways to understand and treat medical conditions.]]></description><content:encoded><![CDATA[<p class="">Does healthcare have anything to learn from military strategies?</p><p class="">In his book “<a href="https://www.amazon.com/Bomber-Mafia-Temptation-Longest-Second/dp/B091J2CP5X" target="_blank">The Bomber Mafia</a>,” Malcolm Gladwell describes how a group of World War II innovators sought to change wartime bombing tactics from broad-based destruction strategies to precision targeting of an adversary’s production capabilities. It is a great book that nicely illustrates how difficult it is to drive a paradigm shift, even one that saves lives.</p><p class="">The health and life sciences industries are facing a <a href="https://www.creoinc.net/smarter-medicine-white-paper/" target="_blank">similar paradigm shift</a> today. Traditional methods of developing novel therapies have relied on population-based approaches to diseases. But <a href="https://en.wikipedia.org/wiki/Fourth_Industrial_Revolution" target="_blank">Industry 4.0 capabilities</a> – artificial intelligence (AI), genomic sciences, Internet of Things (IoT) devices, cloud technologies, and more – are unlocking new opportunities for how we understand and treat medical conditions.</p><h2>The Increasing Role of Precision</h2><p class="">Gladwell’s book emphasizes innovations in <strong>precision</strong> for shifting the existing bombing paradigm. Today, the medical ecosystem sits at a precision transformation as well. Three forces – growing data volumes, better computational technologies, and advancements in biological sciences – are creating a rich landscape of innovation opportunities. Consider two examples:</p><ul data-rte-list="default"><li><p class=""><a href="https://www.fda.gov/medical-devices/software-medical-device-samd/artificial-intelligence-and-machine-learning-aiml-enabled-medical-devices" target="_blank">Over 340 products</a> that include AI have been reviewed by the FDA, and hundreds more are already queued for review. These solutions all have one thing in common: leveraging data and AI algorithms to increase the precision of diagnostics and therapies. The market for analytics like this in health and life sciences is growing at over 25% a year, and RBC Capital has <a href="https://www.rbccm.com/en/gib/healthcare/episode/the_healthcare_data_explosion" target="_blank">estimated</a> that 30% of the world’s data is now related to healthcare — growing faster than any other industry.</p></li></ul><ul data-rte-list="default"><li><p class="">The COVID-19 vaccines you may have received from Pfizer-BioNTech were based on synthetic mRNA technologies using <a href="https://publichealth.jhu.edu/2021/the-long-history-of-mrna-vaccines" target="_blank">lipid nanoparticles</a>. The nanotechnology – operating at a whole new level of precision scale – was critical to being able to deliver mRNA un-degraded into cell proteins. This precision-based approach is now being used for <a href="https://www.pfizer.com/news/articles/the_next_frontier_of_vaccine_innovation" target="_blank">new treatments</a> for influenza, shingles, and many other diseases.</p></li></ul><p class="">So what does a transition towards greater precision mean for health and life sciences companies? We see three broad imperatives for organizations looking to lead in a world of precision medicine and data-driven operations.</p><h3>1. Corporate Strategy &amp; AI</h3><p class="">Business strategies need to reflect that <strong>health and life sciences businesses are competing in an information economy</strong>.<strong> </strong>AI is changing: </p><ul data-rte-list="default"><li><p class="">WHAT companies provide (e.g., targeted drugs; new regenerative medicine applications with 3D printed organs and prosthetics; novel cell and gene therapies based on gene editing; computationally designed antibodies; AI-driven diagnostics; evidence-based therapies; medical devices with embedded AI).</p></li></ul><ul data-rte-list="default"><li><p class="">HOW they provide them (e.g., novel data collection and curation; synthetic research arms; AI-driven drug design; predictive disease modeling; AI-based clinical decision support; automated diagnostics; remote patient monitoring; AI-assisted robotic surgery; clinical trial recruitment modeling and prediction; brain-computer interfaces supporting rehabilitation; new options for driving health outcomes, quality, safety, profitability, and patient engagement).</p></li></ul><p class="">Since precision insights are fueled by excellent data, these business strategies are increasingly reflecting digital transformation as a priority. For example, data-driven companies take process automation seriously – it not only improves productivity and quality, but it also standardizes data collection. Building on this, market leaders are pursuing:</p><ul data-rte-list="default"><li><p class="">Strategies and performance management plans that explicitly identify and measure how precision operations drive competitive differentiation and growth.</p></li></ul><ul data-rte-list="default"><li><p class="">Product and service offerings (e.g., product management roadmaps) that incorporate precision insights as key drivers of value creation with customers.</p></li></ul><ul data-rte-list="default"><li><p class="">Organizational development plans that identify a) key business process automation opportunities, and b) people development plans covering the breadth of skills needed for a data-driven operating posture.</p></li></ul><ul data-rte-list="default"><li><p class="">Data strategy and governance programs that address growing data volumes, depth, timeliness, quality, transparency, and actionability.</p></li></ul><h3>2. Technology &amp; AI</h3><p class=""><strong>Technology roadmaps are critical.</strong> Every health and life sciences organization needs a well-conceived enterprise architecture that can support a data-driven operating model and unique market value: </p><ul data-rte-list="default"><li><p class="">Developing clear “as-is” and “to-be” enterprise architecture designs that align to the business strategy and unique precision characteristics of the organization</p></li></ul><ul data-rte-list="default"><li><p class="">Accommodating AI models that span both central computing architectures and autonomous edge-computing deployments (e.g., hardware chips optimized for federated AI computations and low-power consumption).</p></li></ul><ul data-rte-list="default"><li><p class="">Rapidly integrating new data sources and an increasing collection of interconnected devices that need to exchange data across networks and systems.</p></li></ul><ul data-rte-list="default"><li><p class="">Establishing and managing great “sources of truth” for the organization that provide confidence and trust in the organization’s data assets.</p></li></ul><ul data-rte-list="default"><li><p class="">Effectively deploying analytical insights into line-of-business settings, including no-code / low-code development solutions.</p></li></ul><h3>3. Leadership, Culture &amp; AI</h3><p class=""><strong>Leaders cultivate data-driven cultures.</strong> Despite talented people with good intentions, company cultures do not always promote data-driven work. For leaders looking to leverage AI and data as a differentiator, focus areas should include:</p><ul data-rte-list="default"><li><p class="">Assessing existing barriers to data-driven approaches and insights within the organization’s current culture.</p></li><li><p class="">Creating increased transparency and employee empowerment around data.</p></li><li><p class="">Evoking discussion and collaboration across business, clinical and technology leaders around how analytical insights – designs, simulations, decision support recommendations – can best be incorporated into operations.</p></li><li><p class="">Evaluating and monitoring the safety, efficacy, and appropriateness of AI-driven insights, especially when analytical models may be sourced from 3rd parties.</p></li><li><p class="">Encouraging and rewarding innovation – even failures – to demonstrate the value being placed on discovering new paradigms for the organization.</p></li></ul><p class="">Of course, “precision” is not the only trend driving medical innovation today. And technologies such as CRISPR are demonstrating how broad-based innovations can be applied with high precision as well. But AI’s role is clear: in the war against disease, the soldiers, their military units, and their smart weapons are all increasingly leveraging AI.</p>]]></content:encoded><media:content height="1029" isDefault="true" medium="image" type="image/jpeg" url="https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/1733169214032-9J6MMCG05SDJ70QM1ABI/unsplash-image-sjq4B6PLPug.jpg?format=1500w" width="1500"><media:title type="plain">How AI is Reinventing Medical Innovation</media:title></media:content></item><item><title>Building a Data-Driven Business Strategy</title><category>data</category><dc:creator>Jason Burke</dc:creator><pubDate>Mon, 20 Jul 2020 19:05:00 +0000</pubDate><link>https://www.jasonburke.online/blog/building-a-data-driven-business-strategy</link><guid isPermaLink="false">674a65c6ba0c721888a63309:674a6885532ea52b2bdb097c:674e12fc3ea64a5455267e51</guid><description><![CDATA[Life sciences and healthcare leaders today continue to face a conundrum: 
they want data-driven businesses, but they don’t trust their data or 
insights.]]></description><content:encoded><![CDATA[<p class="">Life sciences and healthcare leaders today continue to face a conundrum.&nbsp; On the one hand, it is clear that data sciences and analytics such as AI can help create a more data-driven, affordable and efficient product development and care delivery ecosystem: design tailored, safe, and effective therapies for rare diseases; streamline drug development; automate high-quality lab and diagnostic services; predict patient demand and care access needs; model disease progression; and countless other use cases.&nbsp; On the other hand, numerous surveys of senior executives by&nbsp;<a href="https://hbr.org/2022/03/overcoming-the-c-suites-distrust-of-ai" target="_blank">Deloitte, KPMG, SAS</a>,&nbsp;<a href="https://www.fiercehealthcare.com/tech/majority-healthcare-executives-don-t-trust-their-organization-s-data-survey-finds" target="_blank">InterSystems</a>, and&nbsp;<a href="https://betanews.com/2021/05/07/enterprise-executives-dont-trust-data/" target="_blank">others</a>&nbsp;have consistently shown that most business leaders do not trust their data and analytics.&nbsp; Beyond the overwhelming number of executives – typically over two-thirds of respondents – expressing this distrust, it’s important to note how persistent this problem has been.</p><h3>A Brief History of Untrusted Insights</h3><p class="">Back in 2006-2007, Thomas Davenport wrote an HBR article and&nbsp;<a href="https://www.amazon.com/Competing-Analytics-New-Science-Winning/dp/1422103323" target="_blank">similarly titled book</a>&nbsp;called&nbsp;<a href="https://hbr.org/2006/01/competing-on-analytics" target="_blank">Competing on Analytics</a>&nbsp;(disclosure: he also generously wrote the foreward to&nbsp;<a href="https://www.amazon.com/Health-Analytics-Gaining-Insights-Transform/dp/1118383044" target="_blank">my book</a>).&nbsp; His works captured an idea that had been gradually emerging since ERP, CRM, and business intelligence systems gained traction in the late 1990s and early 2000s: data and analytics were becoming competitive business competencies, and market leaders will be data-driven.&nbsp; As Davenport shared, the early experiences of organizations on the data-driven path showed that trustworthy data and insights were a problem:</p><p class=""><em>“Executives are aware of the [data quality] problem:</em><br><em>in a survey of the challenges organizations face in developing a business intelligence capability,</em><br><em>data quality was second only to budget constraints.”</em></p><p class="">– Davenport, T (2007).&nbsp;&nbsp;<em>Competing on Analytics</em>,<br>p. 163, Harvard Business School Press.</p><p class="">Over 15 years later, organizations are still struggling to leverage trusted data to become data-driven.&nbsp; And in many ways, the stakes are even greater today.&nbsp; The volume of data we are generating continues to grow exponentially.&nbsp; Regulatory policies related to data privacy, protection, and consumer rights have introduced additional obligations on organizations.&nbsp; And the widespread availability of both data and software tools has dramatically lowered the barriers to entry for competitors, increasing the pressure on business leaders to “figure this out.”</p><p class="">As a result, many executives find themselves in the middle of a self-fulfilling prophecy: when leaders avoid data sciences investments because either the data is not trustworthy or the organization is not seen to be ready for better analytics, they guarantee the data remains untrusted and the organization remains unprepared.&nbsp; As Robert Frost might say, the only way out is through.</p><h3>Finding Value in Data-Driven Insights</h3><p class="">The current fascination, fear, and hype associated with artificial intelligence have raised executive awareness about the opportunity for using data and analytics in new and powerful ways.&nbsp; Leaders and their boards are faced with a fundamental question: what potential business value can be unlocked through investments in data sciences and analytics capabilities?</p><p class="">Experience has shown that building data sciences capabilities in the absence of a clear relationship to its expected business value is an excellent way to waste money.&nbsp; But when those investments are targeted – when the scope of work is focused on specific business questions, workflows, and insights that generate business value – the probability of success grows considerably.</p><p class="">Within life sciences and healthcare, data and analytics are commonly associated with several types of business value:</p><p class=""><strong>1. Better decisions.</strong>&nbsp;One of the most important benefits that data and analytics offer is raising the quality of decisions that need to be made.&nbsp; In our industry, data-driven decisions run a spectrum of concerns including:</p><ul data-rte-list="default"><li><p class="">Clinical / scientific questions: issues related to patients, treatments, and research.</p></li><li><p class="">Financial questions: issues related to revenue, costs, pricing, and reimbursements.</p></li><li><p class="">Administrative questions: issues related to customers, employees, projects, processes, facilities, and supplies.</p></li><li><p class="">Behavioral questions: issues related to activities, preferences and tendencies.</p></li></ul><p class=""><strong>2. Faster actions with lower costs</strong>.&nbsp;Many insights are associated with actions that need to be taken – change study criteria, add manufacturing capacity, conduct a deeper quality review.&nbsp; When these actions are taken sooner, organizations experience productivity gains.&nbsp; In addition, as many insights need to be repeatedly created by different organizational teams, efforts that produce reusable data and analytical assets – effective data curation capabilities, for example – allow all teams to reduce repetitive, error-prone, and time-consuming data and analysis tasks and their costs.&nbsp; And finally, some data-driven actions can also be automated, streamlining and optimizing the use of resources for more complex tasks.</p><p class=""><strong>3. Lower risks</strong>. Data and analytics can be especially useful for identifying emerging risks.&nbsp; Is a patient’s condition declining?&nbsp; Is study enrollment dropping below targets?&nbsp; Does the current commercial pipeline support the future revenue goals?&nbsp; A strong data and analytics strategy enables leaders to detect risks that compromise the most important aspects of business operations and patient care.</p><p class=""><strong>4. Better products and customer experiences.</strong>&nbsp;Many organizations are discovering that a well-conceived data and analytics strategy can be used to create highly differentiated customer experiences.&nbsp; Customers value on-demand access to current information (i.e., what is the status), transparency with business processes (i.e., what steps have been completed), and streamlined ways of engaging with their partners (i.e., where is all of the information).&nbsp; By bringing disparate information together in ways that are easy to consume, organizations can create compelling products and services that drive higher customer lifetime value.</p><p class=""><strong>5. Stronger security and compliance</strong>.&nbsp;Well-developed data and analytical strategies also pay dividends in regulatory and security management.&nbsp; When organizations have more disciplined approaches to managing data and insights, it becomes much easier to manage security (i.e., who sees what) and regulatory (e.g., how are controls managed) obligations.&nbsp; In addition, data and analytics can also be used to detect patterns of electronic behaviors indicative of security or regulatory lapses (e.g., anomalous system access, excessive data retrieval).</p><p class="">So if these are the types of value commonly available, how do you identify which investments to prioritize?</p><h3>Aligning Corporate and Insight Strategies</h3><p class="">Historically, investment decisions regarding data and analytics were pursued relatively tactically primarily through operational reports (i.e., from line-of-business systems) and data extracts (e.g., for manipulation in Microsoft Excel).&nbsp; Today, we know those approaches do not scale: there are simply too many systems, data, quality issues, access control obligations, reports, and data consumers to serve.&nbsp; In addition, organizations need insights for more than simply good operational oversight – insights serve a more strategic function for the business.</p><p class="">Excellent business strategies provide unambiguous focus – a lens for continuously discerning the “important” from the “good”.&nbsp; The same disciplines we use in strategic planning can be incredibly useful in helping organizations plan investment strategies in data and analytics.&nbsp; We can use strategic planning artifacts – goals, strategic imperatives, and corporate capabilities – to characterize and prioritize the internal (i.e., employee-facing) and external (i.e., visible to customers and partners) insights that best support the corporate strategy and growth goals.  Ideally, we want to focus on investments that accelerate value creation related to our strategy.&nbsp; Questions we might want to explore include:</p><ol data-rte-list="default"><li><p class="">Based on our business model and market differentiation, what data and insights are most crucial in helping us make the best decisions about our organization?</p></li><li><p class="">What data and analytical capabilities would make us more valuable and attractive to investors in this market space?</p></li><li><p class="">What insights would help us know whether our corporate strategy is successful?</p></li><li><p class="">What time-sensitive metrics would better enable us to capture opportunities for operational improvements as we execute our strategy?</p></li><li><p class="">How could data and insights facilitate a totally different customer experience with our products and services?</p></li><li><p class="">How could we use data to detect and manage emerging risks for us and our customers?</p></li><li><p class="">How will competitors likely use data, analytics, and insights to win business?</p></li><li><p class="">How do we know if we are meeting our regulatory obligations regarding data privacy, security, and protections?</p></li></ol><p class="">Grounded in the business value opportunities mentioned above, these questions (and others like them) serve to orient leadership thinking about data and analytics as both foundational business assets and key organizational competencies.&nbsp; They also help to triage insights that contribute to&nbsp;building trusted, data-driven cultures.</p><p class="">So what role are data and analytics playing in your business strategy?</p>]]></content:encoded><media:content height="1000" isDefault="true" medium="image" type="image/jpeg" url="https://images.squarespace-cdn.com/content/v1/674a65c6ba0c721888a63309/1733170838198-GV3QIE6SKV155F769WAJ/unsplash-image-fzOITuS1DIQ.jpg?format=1500w" width="1500"><media:title type="plain">Building a Data-Driven Business Strategy</media:title></media:content></item></channel></rss>